6,640 results
Search Results
52. Recommender Systems for Outdoor Adventure Tourism Sports: Hiking, Running and Climbing.
- Author
-
Ivanova, Iustina and Wald, Mike
- Subjects
HIKING ,ADVENTURE tourism ,INTERNATIONAL visitors ,RECOMMENDER systems ,DEEP learning ,MACHINE learning ,ARTIFICIAL intelligence - Abstract
Adventure tourism is a popular and growing segment within the tourism industry that involves, but is not limited to, hiking, running, and climbing activities. These activities attract investment from foreign travelers interested in practicing sports while exploring other countries. As a result, many software companies started developing Artificial Intelligence solutions to enhance tourists' outdoor adventure experience. One of the leading technologies in this field is recommender systems, which provide personalized recommendations to tourists based on their preferences. While this topic is actively being researched in some sports (running and hiking), other adventure sports disciplines have yet to be fully explored. To standardize the development of intelligence-based recommender systems, we conducted a systematic literature review on more than a thousand scientific papers published in decision support system applications in three outdoor adventure sports, such as running, hiking, and sport climbing. Hence, the main focus of this work is, firstly, to summarize the state-of-the-art methods and techniques being researched and developed by scientists in recommender systems in adventure tourism, secondly, to provide a unified methodology for software solutions designed in this domain, and thirdly, to give further insights into open possibilities in this topic. This literature survey serves as a unified framework for the future development of technologies in adventure tourism. Moreover, this paper seeks to guide the development of more effective and personalized recommendation systems. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
53. Modified receiver architecture in software-defined radio for real-time modulation classification.
- Author
-
Le, Quoc Nam, Huynh, Tan Quoc, Ta, Hien Quang, Tan, Phuoc Vo, and Nguyen, Lap Luat
- Subjects
SOFTWARE radio ,ELECTRONIC modulation ,ARTIFICIAL neural networks ,DEEP learning ,SPECTRUM allocation ,TELECOMMUNICATION systems - Abstract
Automatic modulation classification (AMC) is an important process for future communication systems with prominent applications from spectrum management, and secure communication, to cognitive radio. The requirement for an efficient AMC classifier is due to its capability in blind modulation recognition, which is a difficult task in real scenarios where the limitations of traditional hardware and the complexity of channel impairments are involved. Therefore, this paper proposes a complete real-time AMC system based on software-defined radio and deep learning architecture. The system demodulation performance is verified through simulations and real channel impairment conditions to ensure reliability. With at most 6 times reduced number of parameters, two proposed models convolutional long short-term memory deep neural network and residual long short-term memory neural network also show a general improvement in classification accuracy compared with reference studies. The performance of these models at real-time AMC is tested with suitable processing time for practical applications. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
54. Semantic segmentation of point clouds of ancient buildings based on weak supervision.
- Author
-
Zhao, Jianghong, Yu, Haiquan, Hua, Xinnan, Wang, Xin, Yang, Jia, Zhao, Jifu, and Xu, Ailin
- Subjects
POINT cloud ,ANCIENT architecture ,HISTORIC buildings ,BUILDING information modeling ,PROFESSIONALISM - Abstract
Semantic segmentation of point clouds of ancient buildings plays an important role in Historical Building Information Modelling (HBIM). As the annotation task of point cloud of ancient architecture is characterised by strong professionalism and large workload, which greatly restricts the application of point cloud semantic segmentation technology in the field of ancient architecture, therefore, this paper launches a research on the semantic segmentation method of point cloud of ancient architecture based on weak supervision. Aiming at the problem of small differences between classes of ancient architectural components, this paper introduces a self-attention mechanism, which can effectively distinguish similar components in the neighbourhood. Moreover, this paper explores the insufficiency of positional encoding in baseline and constructs a high-precision point cloud semantic segmentation network model for ancient buildings—Semantic Query Network based on Dual Local Attention (SQN-DLA). Using only 0.1% of the annotations in our homemade dataset and the Architectural Cultural Heritage (ArCH) dataset, the mean Intersection over Union (mIoU) reaches 66.02% and 58.03%, respectively, which is an improvement of 3.51% and 3.91%, respectively, compared to the baseline. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
55. An appearance quality classification method for Auricularia auricula based on deep learning.
- Author
-
Li, Yang, Hu, Jiajun, Wu, Haiyun, Wei, Yong, Shan, Huiyong, Song, Xin, Hua, Xiuping, Xu, Wei, and Jiang, Yongcheng
- Subjects
DEEP learning ,PRIMROSES ,CONVOLUTIONAL neural networks ,DATA mining ,FEATURE extraction ,CLASSIFICATION - Abstract
The intelligent appearance quality classification method for Auricularia auricula is of great significance to promote this industry. This paper proposes an appearance quality classification method for Auricularia auricula based on the improved Faster Region-based Convolutional Neural Networks (improved Faster RCNN) framework. The original Faster RCNN is improved by establishing a multiscale feature fusion detection model to improve the accuracy and real-time performance of the model. The multiscale feature fusion detection model makes full use of shallow feature information to complete target detection. It fuses shallow features with rich detailed information with deep features rich in strong semantic information. Since the fusion algorithm directly uses the existing information of the feature extraction network, there is no additional calculation. The fused features contain more original detailed feature information. Therefore, the improved Faster RCNN can improve the final detection rate without sacrificing speed. By comparing with the original Faster RCNN model, the mean average precision (mAP) of the improved Faster RCNN is increased by 2.13%. The average precision (AP) of the first-level Auricularia auricula is almost unchanged at a high level. The AP of the second-level Auricularia auricula is increased by nearly 5%. And the third-level Auricularia auricula AP is increased by 1%. The improved Faster RCNN improves the frames per second from 6.81 of the original Faster RCNN to 13.5. Meanwhile, the influence of complex environment and image resolution on the Auricularia auricula detection is explored. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
56. SIGNIFICANCE deep learning based platform to fight illicit trafficking of Cultural Heritage goods.
- Author
-
Malinverni, Eva Savina, Abate, Dante, Agapiou, Antonia, Stefano, Francesco Di, Felicetti, Andrea, Paolanti, Marina, Pierdicca, Roberto, and Zingaretti, Primo
- Subjects
DEEP learning ,CULTURAL property ,ARTIFICIAL intelligence ,LAW enforcement agencies ,DARKNETS (File sharing) ,SOCIAL media ,ONTOLOGIES (Information retrieval) - Abstract
The illicit traffic of cultural goods remains a persistent global challenge, despite the proliferation of comprehensive legislative frameworks developed to address and prevent cultural property crimes. Online platforms, especially social media and e-commerce, have facilitated illegal trade and pose significant challenges for law enforcement agencies. To address this issue, the European project SIGNIFICANCE was born, with the aim of combating illicit traffic of Cultural Heritage (CH) goods. This paper presents the outcomes of the project, introducing a user-friendly platform that employs Artificial Intelligence (AI) and Deep learning (DL) to prevent and combat illicit activities. The platform enables authorities to identify, track, and block illegal activities in the online domain, thereby aiding successful prosecutions of criminal networks. Moreover, it incorporates an ontology-based approach, providing comprehensive information on the cultural significance, provenance, and legal status of identified artefacts. This enables users to access valuable contextual information during the scraping and classification phases, facilitating informed decision-making and targeted actions. To accomplish these objectives, computationally intensive tasks are executed on the HPC CyClone infrastructure, optimizing computing resources, time, and cost efficiency. Notably, the infrastructure supports algorithm modelling and training, as well as web, dark web and social media scraping and data classification. Preliminary results indicate a 10–15% increase in the identification of illicit artifacts, demonstrating the platform's effectiveness in enhancing law enforcement capabilities. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
57. A Beginner's Guide to Artificial Intelligence for Ophthalmologists.
- Author
-
Kang, Daohuan, Wu, Hongkang, Yuan, Lu, Shi, Yu, Jin, Kai, and Grzybowski, Andrzej
- Subjects
ARTIFICIAL intelligence ,OPHTHALMOLOGISTS ,DEEP learning ,DIAGNOSTIC errors ,BALANCE disorders ,TREATMENT effectiveness ,ADRENAL insufficiency - Abstract
The integration of artificial intelligence (AI) in ophthalmology has promoted the development of the discipline, offering opportunities for enhancing diagnostic accuracy, patient care, and treatment outcomes. This paper aims to provide a foundational understanding of AI applications in ophthalmology, with a focus on interpreting studies related to AI-driven diagnostics. The core of our discussion is to explore various AI methods, including deep learning (DL) frameworks for detecting and quantifying ophthalmic features in imaging data, as well as using transfer learning for effective model training in limited datasets. The paper highlights the importance of high-quality, diverse datasets for training AI models and the need for transparent reporting of methodologies to ensure reproducibility and reliability in AI studies. Furthermore, we address the clinical implications of AI diagnostics, emphasizing the balance between minimizing false negatives to avoid missed diagnoses and reducing false positives to prevent unnecessary interventions. The paper also discusses the ethical considerations and potential biases in AI models, underscoring the importance of continuous monitoring and improvement of AI systems in clinical settings. In conclusion, this paper serves as a primer for ophthalmologists seeking to understand the basics of AI in their field, guiding them through the critical aspects of interpreting AI studies and the practical considerations for integrating AI into clinical practice. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
58. A survey on applications of reinforcement learning in spatial resource allocation.
- Author
-
Zhang, Di, Wang, Moyang, Mango, Joseph, Li, Xiang, and Xu, Xianrui
- Subjects
REINFORCEMENT learning ,RESOURCE allocation ,DEEP learning - Abstract
The challenge of spatial resource allocation is pervasive across various domains such as transportation, industry, and daily life. As the scale of real-world issues continues to expand and demands for real-time solutions increase, traditional algorithms face significant computational pressures, struggling to achieve optimal efficiency and real-time capabilities. In recent years, with the escalating computational power of computers, the remarkable achievements of reinforcement learning in domains like Go and robotics have demonstrated its robust learning and sequential decision-making capabilities. Given these advancements, there has been a surge in novel methods employing reinforcement learning to tackle spatial resource allocation problems. These methods exhibit advantages such as rapid solution convergence and strong model generalization abilities, offering a new perspective on resolving spatial resource allocation problems. Despite the progress, reinforcement learning still faces hurdles when it comes to spatial resource allocation. There remains a gap in its ability to fully grasp the diversity and intricacy of real-world resources. The environmental models used in reinforcement learning may not always capture the spatial dynamics accurately. Moreover, in situations laden with strict and numerous constraints, reinforcement learning can sometimes fall short in offering feasible strategies. Consequently, this paper is dedicated to summarizing and reviewing current theoretical approaches and practical research that utilize reinforcement learning to address issues pertaining to spatial resource allocation. In addition, the paper accentuates several unresolved challenges that urgently necessitate future focus and exploration within this realm and proposes viable approaches for these challenges. This research furnishes valuable insights that may assist scholars in gaining a more nuanced understanding of the problems, opportunities, and potential directions concerning the application of reinforcement learning in spatial resource allocation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
59. ECG autoencoder based on low-rank attention.
- Author
-
Zhang, Shilin, Fang, Yixian, and Ren, Yuwei
- Subjects
ARTIFICIAL neural networks ,SINGULAR value decomposition ,ELECTROCARDIOGRAPHY ,DISEASE prevalence ,DEEP learning - Abstract
The prevalence of cardiovascular disease (CVD) has surged in recent years, making it the foremost cause of mortality among humans. The Electrocardiogram (ECG), being one of the pivotal diagnostic tools for cardiovascular diseases, is increasingly gaining prominence in the field of machine learning. However, prevailing neural network models frequently disregard the spatial dimension features inherent in ECG signals. In this paper, we propose an ECG autoencoder network architecture incorporating low-rank attention (LRA-autoencoder). It is designed to capture potential spatial features of ECG signals by interpreting the signals from a spatial perspective and extracting correlations between different signal points. Additionally, the low-rank attention block (LRA-block) obtains spatial features of electrocardiogram signals through singular value decomposition, and then assigns these spatial features as weights to the electrocardiogram signals, thereby enhancing the differentiation of features among different categories. Finally, we utilize the ResNet-18 network classifier to assess the performance of the LRA-autoencoder on both the MIT-BIH Arrhythmia and PhysioNet Challenge 2017 datasets. The experimental results reveal that the proposed method demonstrates superior classification performance. The mean accuracy on the MIT-BIH Arrhythmia dataset is as high as 0.997, and the mean accuracy and F 1 -score on the PhysioNet Challenge 2017 dataset are 0.850 and 0.843. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
60. A robust approach to authorship verification using siamese deep learning: application in phishing email detection.
- Author
-
Remmide, Mohamed Abdelkarim, Boumahdi, Fatima, Ammar Aouchiche, Imane Rebeh, Guendouz, Amina, and Boustia, Narhimene
- Abstract
Given the rapid and significant increase in email data, it is crucial for both individuals and organisations to prioritise the implementation of strong cybersecurity measures to combat attacks such as phishing emails. While continuous research has been made to find the most efficient approach to combating phishing emails, cybercriminals on the other hand continue to be determined, and their techniques become more and more sophisticated. In this paper, we present a novel approach for email classification using a Siamese deep learning network in order to verify authorship. We used the writing style and behavioural characteristics of the authors in this approach to identify the true source of emails and classify them as phishing, legitimate, harassment, or suspicious. This model was evaluated on the SeFACED dataset, where it received an accuracy of 90,12%, showcasing its efficacy in classifying emails, which enhanced email security and contributed to a safer online environment. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
61. HSS: enhancing IoT malicious traffic classification leveraging hybrid sampling strategy.
- Author
-
Luo, Yuantu, Tao, Jun, Zhu, Yuehao, and Xu, Yifan
- Subjects
COMPUTER network traffic ,INTERNET of things ,STATISTICAL sampling ,AUTOMATIC classification ,DEEP learning ,CLASSIFICATION - Abstract
Using deep learning models to deal with the classification tasks in network traffic offers a new approach to address the imbalanced Internet of Things malicious traffic classification problems. However, the employment difficulty of these models may be immense due to their high resource consumption and inadequate interpretability. Fortunately, the effectiveness of sampling methods based on the statistical principles in imbalance data distribution indicates the path. In this paper, we address these challenges by proposing a hybrid sampling method, termed HSS, which integrates undersampling and oversampling techniques. Our approach not only mitigates the imbalance in malicious traffic but also fine-tunes the sampling threshold to optimize performance, as substantiated through validation tests. Employed across three distinct classification tasks, this method furnishes simplified yet representative samples, enhancing the baseline models' classification capabilities by a minimum of 6.02% and a maximum of 182.66%. Moreover, it notably reduces resource consumption, with sample numbers diminishing to a ratio of at least 83.53%. This investigation serves as a foundation, demonstrating the efficacy of HSS in bolstering security measures in IoT networks, potentially guiding the development of more adept and resource-efficient solutions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
62. Volatility Spillovers and Contagion During Major Crises: An Early Warning Approach Based on a Deep Learning Model.
- Author
-
Sahiner, Mehmet
- Subjects
PORTFOLIO diversification ,DEEP learning ,FINANCIAL stress ,GLOBAL Financial Crisis, 2008-2009 ,EMERGING markets ,COVID-19 pandemic - Abstract
This paper contributes to the ongoing debate on the nature and characteristics of the volatility transmission channels of major crash events in international stock markets between 03 July 1997 and 09 March 2021. Using dynamic conditional correlations (DCC) for conditional correlations and volatility clustering, GARCH-BEKK for the direction of transmission of disturbances, and the Diebold-Yilmaz spillover index for the level of volatility contagion, the paper finds that the climbs in external shock transmissions have long-lasting impacts in domestic markets due to the contagion effect during crisis periods. The findings also reveal that the heavier magnitude of financial stress is transmitted between Asian countries via the Hong Kong stock market. Additionally, the degree of volatility spillovers between advanced and emerging equity markets is smaller compared to the pure spillovers between advanced markets or emerging markets, offering a window of opportunity for international market participants in terms of portfolio diversification and risk management applications. Furthermore, the study introduces a novel early warning system created by integrating DCC correlations with a state-of-the-art deep learning model to predict the global financial crisis and COVID-19 crisis. The experimental analysis of long short-term memory network finds evidence of contagion risk by verifying bursts in volatility spillovers and generating signals with high accuracy before the 12-month crisis period. This provides supplementary information that contributes to the decision-making process of practitioners, as well as offering indicative evidence that facilitates the assessment of market vulnerability for policymakers. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
63. Ancient mural dynasty recognition algorithm based on a neural network architecture search.
- Author
-
Cao, Jianfang, Jin, Mengyan, Tian, Yun, Cao, Zhen, and Peng, Cunhe
- Subjects
ARTIFICIAL neural networks ,MURAL art ,RECOGNITION (Psychology) ,IMAGE recognition (Computer vision) ,ALGORITHMS - Abstract
A neural network model needs to be manually designed for ancient mural dynasty recognition, and this paper proposes an ancient mural dynasty recognition algorithm that is based on a neural architecture search (NAS). First, the structural edge information of mural images is extracted for use by the neural network model in recognizing mural missions. Second, an NAS algorithm that is based on contrast selection (CS) simplifies the architecture search to an incremental CS and then searches for the optimal network architecture on the mural dataset. Finally, the identified optimal network architecture is used for training and testing to complete the mural dynasty recognition task. The results show that the top accuracy of the proposed method on the mural dataset is 88.10%, the recall rate is 87.52%, and the precision rate is 87.69%. Each evaluation index used by the neural network model is superior to that of classical network models such as AlexNet and ResNet-50. Compared with NAS methods such as ASNG and MIGO, the accuracy of mural dynasty recognition is higher by an average of 4.27% when using the proposed method. The proposed method is verified on CIFAR-10, CIFAR-100, ImageNet16-120 and other datasets and achieves a good recognition accuracy in the NAS-bench-201 search space, which averages 93.26%, 70.73% and 45.34%, respectively, on the abovementioned datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
64. STAM-LSGRU: a spatiotemporal radar echo extrapolation algorithm with edge computing for short-term forecasting.
- Author
-
Cheng, Hailang, Cui, Mengmeng, and Shi, Yuzhe
- Subjects
EDGE computing ,RADAR ,MOBILE computing ,ALGORITHMS ,WEATHER forecasting - Abstract
With the advent of Mobile Edge Computing (MEC), shifting data processing from cloud centers to the network edge presents an advanced computational paradigm for addressing latency-sensitive applications. Specifically, in radar systems, the real-time processing and prediction of radar echo data pose significant challenges in dynamic and resource-constrained environments. MEC, by processing data near its source, not only significantly reduces communication latency and enhances bandwidth utilization but also diminishes the necessity of transmitting large volumes of data to the cloud, which is crucial for improving the timeliness and efficiency of radar data processing. To meet this demand, this paper proposes a model that integrates a spatiotemporal Attention Module (STAM) with a Long Short-Term Memory Gated Recurrent Unit (ST-ConvLSGRU) to enhance the accuracy of radar echo prediction while leveraging the advantages of MEC. STAM, by extending the spatiotemporal receptive field of the prediction units, effectively captures key inter-frame motion information, while optimizations to the convolutional structure and loss function further boost the model's predictive performance. Experimental results demonstrate that our approach significantly improves the accuracy of short-term weather forecasting in a mobile edge computing environment, showcasing an efficient and practical solution for processing radar echo data under dynamic, resource-limited conditions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
65. The analysis of ecological security and tourist satisfaction of ice-and-snow tourism under deep learning and the Internet of Things.
- Author
-
Zhang, Baiju
- Subjects
ENVIRONMENTAL security ,SATISFACTION ,INTERNET of things ,RECURRENT neural networks ,DEEP learning ,TOURISM ,TOURISTS - Abstract
This paper aims to propose a prediction method based on Deep Learning (DL) and Internet of Things (IoT) technology, focusing on the ecological security and tourist satisfaction of Ice-and-Snow Tourism (IST) to solve practical problems in this field. Accurate predictions of ecological security and tourist satisfaction in IST have been achieved by collecting and analyzing environment and tourist behavior data and combining with DL models, such as convolutional and recurrent neural networks. The experimental results show that the proposed method has significant advantages in performance indicators, such as accuracy, F1 score, Mean Squared Error (MSE), and correlation coefficient. Compared to other similar methods, the method proposed improves accuracy by 3.2%, F1 score by 0.03, MSE by 0.006, and correlation coefficient by 0.06. These results emphasize the important role of combining DL with IoT technology in predicting ecological security and tourist satisfaction in IST. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
66. Combining graph neural network with deep reinforcement learning for resource allocation in computing force networks.
- Author
-
Han, Xueying, Xie, Mingxi, Yu, Ke, Huang, Xiaohong, Du, Zongpeng, and Yao, Huijuan
- Abstract
Copyright of Frontiers of Information Technology & Electronic Engineering is the property of Springer Nature and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
67. GARCHNet: Value-at-Risk Forecasting with GARCH Models Based on Neural Networks.
- Author
-
Buczynski, Mateusz and Chlebus, Marcin
- Subjects
STANDARD & Poor's 500 Index ,GARCH model ,VALUE at risk ,MAXIMUM likelihood statistics ,DEEP learning - Abstract
This paper proposes a new GARCH specification that adapts the architecture of a long-term short memory neural network (LSTM). It is shown that classical GARCH models generally give good results in financial modeling, where high volatility can be observed. In particular, their high value is often praised in Value-at-Risk. However, the lack of nonlinear structure in most approaches means that conditional variance is not adequately represented in the model. On the contrary, the recent rapid development of deep learning methods is able to describe any nonlinear relationship in a clear way. We propose GARCHNet, a nonlinear approach to conditional variance that combines LSTM neural networks with maximum likelihood estimators in GARCH. The variance distributions considered in the paper are normal, t and skewed t, but the approach allows extension to other distributions. To evaluate our model, we conducted an empirical study on the logarithmic returns of the WIG 20 (Warsaw Stock Exchange Index), S&P 500 (Standard & Poor's 500) and FTSE 100 (Financial Times Stock Exchange) indices over four different time periods from 2005 to 2021 with different levels of observed volatility. Our results confirm the validity of the solution, but we provide some directions for its further development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
68. Guest Editorial: Advanced Machine Learning Algorithms and Signal Processing.
- Author
-
Manogaran, Gunasekaran, Chilamkurti, Naveen, and Hsu, Ching-Hsien
- Subjects
MACHINE learning ,SIGNAL processing ,FUZZY algorithms ,ARTIFICIAL neural networks ,DEEP learning - Abstract
This special issue of the circuits, systems and signal processing journal focuses on recent advances and improvements in leading-edge integrated machine learning algorithms and signal processing systems. Although signal processing has been studied over several decades, the computer industry is only beginning to understand how signal processing techniques can be well integrated for the development of human-machine interfaces with the advancement of machine learning algorithms. As advances in signal processing tools and machine learning algorithms are becoming more powerful in terms of functionality and communicative capabilities, their contribution to the journal of circuits, system and signal processing is becoming more significant. [Extracted from the article]
- Published
- 2020
- Full Text
- View/download PDF
69. Detection of epileptic seizure in EEG signals using machine learning and deep learning techniques.
- Author
-
Kunekar, Pankaj, Gupta, Mukesh Kumar, and Gaur, Pramod
- Subjects
EPILEPSY ,MACHINE learning ,DEEP learning ,ELECTROENCEPHALOGRAPHY ,FEATURE extraction ,DIAGNOSIS of epilepsy - Abstract
Around 50 million individuals worldwide suffer from epilepsy, a chronic, non-communicable brain disorder. Several screening methods, including electroencephalography, have been proposed to identify epileptic episodes. EEG data, which are frequently utilised to enhance epilepsy analysis, offer essential information on the electrical processes of the brain. Prior to the emergence of deep learning (DL), feature extraction was accomplished by standard machine learning techniques. As a result, they were only as good as the people who made the features by hand. But with DL, both feature extraction and classification are fully automated. These methods have significantly advanced several fields of medicine, including the diagnosis of epilepsy. In this paper, the works focused on automated epileptic seizure detection using ML and DL techniques are presented as well as their comparative analysis is done. The UCI-Epileptic Seizure Recognition dataset is used for training and validation. Some of the conventional ML and DL algorithms are used with a proposed model which uses long short-term memory (LSTM) to find the best approach. Post that comparative analysis is performed on these algorithms to find the best approach for epileptic seizure detection. As a result, the proposed model LSTM gives a validation accuracy of 97% giving the most appropriate and precise result as compared to other mentioned algorithms used in this study. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
70. An efficient EEG signal fading processing framework based on the cognitive limbic system and deep learning.
- Author
-
Wang, Wenlong, Li, Baojiang, Wang, Haiyan, and Wang, Xichao
- Subjects
DEEP learning ,LIMBIC system ,SIGNAL processing ,ELECTROENCEPHALOGRAPHY ,STANDARD deviations ,BRAIN-computer interfaces - Abstract
Non-invasive electroencephalography (EEG) is a technique for monitoring brain activity that is valuable in the diagnosis and study of the brain. However, due to factors such as brain-computer interface (BCI) devices deficiency, dynamic network limitations, and subject issues, EEG signals may fade throughout the entire process from signal generation to acquisition. The fading of EEG signals can cause changes in the data distribution, blur the information and have a negative impact on subsequent research and application. In order to reduce the adverse effects of data fading, this paper proposes a hierarchical bidirectional long-short term memory (LSTM)-Attention network based on cognitive brain limbic system (HBLANet). HBLANet classifies randomly fading signals multiple times by way of a dimensionality-reducing classification algorithm in order to narrow down the feature interval processed by a single neural network. Then the different types of EEG signals acquired from the classification are processed in a more focused manner using a bidirectional LSTM-Attention network. In this paper, the overall performance of the network is greatly improved by decomposing complex signal processing tasks into smaller tasks. The model is evaluated on the EEG-denoisenet dataset and compared with competitive networks such as feature pyramid network (FPN), UNet, MultiResUNet, 1D-ResCNN etc., and the results show that the proposed network achieves better fading processing outcomes. In the overall experiment, the processed EEG signals achieve the relative root mean squared error (RRMSE) value of 0.009, the signal-to-noise ratio (SNR) of 32.78, and the correlation coefficient (CC) of 0.98. Furthermore, the denoising task in the overall experiment achieves even more exciting results, with a SNR of 40.48 and a CC of 0.991 for the processed EEG signal in the case of processing high SNR signals (the range of the SNR is from -2 to 2). Therefore, we believe that the framework has an important reference value for future research on signal quality restoration. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
71. F3l: an automated and secure function-level low-overhead labeled encrypted traffic dataset construction method for IM in Android.
- Author
-
Xu, Keya and Cheng, Guang
- Subjects
MOBILE operating systems ,INSTANT messaging ,DEEP learning ,MACHINE learning ,COMPUTER network security - Abstract
Fine-grained function-level encrypted traffic classification is an essential approach to maintaining network security. Machine learning and deep learning have become mainstream methods to analyze traffic, and labeled dataset construction is the basis. Android occupies a huge share of the mobile operating system market. Instant Messaging (IM) applications are important tools for people communication. But such applications have complex functions which frequently switched, so it is difficult to obtain function-level labels. The existing function-level public datasets in Android are rare and noisy, leading to research stagnation. Most labeled samples are collected with WLAN devices, which cannot exclude the operating system background traffic. At the same time, other datasets need to obtain root permission or use scripts to simulate user behavior. These collecting methods either destroy the security of the mobile device or ignore the real operation features of users with coarse-grained. Previous work (Chen et al. in Appl Sci 12(22):11731, 2022) proposed a one-stop automated encrypted traffic labeled sample collection, construction, and correlation system, A3C, running at the application-level in Android. This paper analyzes the display characteristics of IM and proposes a function-level low-overhead labeled encrypted traffic datasets construction method for Android, F3L. The supplementary method to A3C monitors UI controls and layouts of the Android system in the foreground. It selects the feature fields of attributes of them for different in-app functions to build an in-app function label matching library for target applications and in-app functions. The deviation of timestamp between function invocation and label identification completion is calibrated to cut traffic samples and map them to corresponding labels. Experiments show that the method can match the correct label within 3 s after the user operation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
72. Low-shot learning and class imbalance: a survey.
- Author
-
Billion Polak, Preston, Prusa, Joseph D., and Khoshgoftaar, Taghi M.
- Subjects
LANGUAGE models ,EVIDENCE gaps ,BIG data ,DEEP learning - Abstract
The tasks of few-shot, one-shot, and zero-shot learning—or collectively "low-shot learning" (LSL)—at first glance are quite similar to the long-standing task of class imbalanced learning; specifically, they aim to learn classes for which there is little labeled data available. Motivated by this similarity, we conduct a survey to review the recent literature for works which combine these fields in one of two ways, either addressing the obstacle of class imbalance within a LSL setting, or utilizing LSL techniques or frameworks in order to combat class imbalance within other settings. In our survey of over 60 papers in a wide range of applications from January 2020 to July 2023 (inclusive), we examine and report methodologies and experimental results, find that most works report performance at or above their respective state-of-the-art, and highlight current research gaps which hold potential for future work, especially those involving the use of LSL techniques in imbalanced tasks. To this end, we emphasize the lack of works utilizing LSL approaches based on large language models or semantic data, and works using LSL for big-data imbalanced tasks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
73. Advances and challenges in artificial intelligence text generation.
- Author
-
Li, Bing, Yang, Peng, Sun, Yuankang, Hu, Zhongjian, and Yi, Meng
- Abstract
Copyright of Frontiers of Information Technology & Electronic Engineering is the property of Springer Nature and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
74. Generating adaptation rule-specific neural networks.
- Author
-
Bureš, Tomáš, Hnětynka, Petr, Kruliš, Martin, Plášil, František, Khalyeyev, Danylo, Hahner, Sebastian, Seifermann, Stephan, Walter, Maximilian, and Heinrich, Robert
- Subjects
NEUROPLASTICITY ,DEEP learning ,REINFORCEMENT learning ,MACHINE learning ,PHYSIOLOGICAL adaptation - Abstract
There have been a number of approaches to employ neural networks in self-adaptive systems; in many cases, generic neural networks and deep learning are utilized for this purpose. When this approach is to be applied to improve an adaptation process initially driven by logical adaptation rules, the problem is that (1) these rules represent a significant and tested body of domain knowledge, which may be lost if they are replaced by a neural network, and (2) the learning process is inherently demanding given the black-box nature and the number of weights in generic neural networks to be trained. In this paper, we introduce the rule-specific neural network method that makes it possible to transform the guard of an adaptation rule into a rule-specific neural network, the composition of which is driven by the structure of the logical predicates in the guard. Our experiments confirmed that the black box effect is eliminated, the number of weights is significantly reduced, and much faster learning is achieved whilst the accuracy is preserved. This text is an extended version of the paper presented at the ISOLA 2022 conference (Bureš et al. in Proceedings of ISOLA 2022, Rhodes, Greece, pp. 215–230, 2022). [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
75. A robust energy management approach in two-steps ahead using deep learning BiLSTM prediction model and type-2 fuzzy decision-making controller.
- Author
-
El Bourakadi, Dounia, Ramadan, Hiba, Yahyaouy, Ali, and Boumhidi, Jaouad
- Subjects
DEEP learning ,DECISION making ,PREDICTION models ,STATISTICAL decision making ,ENERGY management ,ELECTRICITY pricing ,ITERATIVE learning control - Abstract
The price prediction is valuable in energy management system (EMS) because it allows making informed decisions and solving the problem of the uncertainty related to the future ignorance based only on the past knowledge. To this goal, we present in this paper a two-steps EMS in order to control the different operations of a micro-grid (MG). In the first step, we exploit the advantages of the Bidirectional Long-Short Term Memory (BiLSTM) deep learning model to predict the next daily electricity price based on time series. In the second step, we use a type-2 fuzzy logic controller to decide which energy source will exploit the excess energy produced or meet the MG need. Real data is used in this paper to test the effectiveness of the proposed EMS whose superiority is proved through the test period. The BiLSTM forecasting model better performs compared to other related algorithms designed to the electricity price prediction. In addition, the proposed decision-making process can reduce the total MG cost and protect the batteries against the deep discharge and maximum charge in order to prolong their lifespan. We expect that this work can contribute to meet the real-world needs in the management of the electrical system. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
76. Visualization analysis of research hotspots on structural topology optimization based on CiteSpace.
- Author
-
Zhong, Yi, Jiang, Xue-tao, Yang, Yong, Xu, Ben-lian, Zhu, Qi-xin, Wang, Lei, and Dong, Xin-feng
- Subjects
STRUCTURAL optimization ,LEVEL set methods ,STRUCTURAL design ,DATA visualization ,DEEP learning - Abstract
Structural topology optimization has gained widespread attention due to more possibilities of innovative structural design. The current research focus/hotspots, application areas, main research scholars, institutions and the countries involved in structural topology optimization are visually presented through clustering and visual analysis based on CiteSpace. The four metric dimensions of the literatures in this paper are as follows: annual quantity of papers and core countries, core authors and co-authors' institutions, hotspots and burst terms, and the highly co-cited papers. The results show the research hotspots in this field are concentrated on keywords such as "level set method", "sensitivity analysis", "homogenization", "genetic algorithm", etc. Regarding the research frontier, "moving morphable component (MMC)", "additive manufacturing (AM)" and "deep learning" are hot topics. In addition, Y. Sui, Z. Kang and O. Sigmund, etc. have high publications. M. Bendsøe and O. Sigmund have high citations. Dalian University of Technology, Technical University of Denmark, etc. are prominent institutions. Moreover, China accounts for more than 34% in the terms of original WOS literatures following by the USA and Australia. This paper could identify structural topology optimization development patterns for the scholars concerned with this field, especially novices, to quickly focus and track the research priorities. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
77. Ternary symmetric fusion network for camouflaged object detection.
- Author
-
Deng, Yangyang, Ma, Jianxin, Li, Yajun, Zhang, Min, and Wang, Li
- Subjects
OBJECT recognition (Computer vision) ,DEEP learning - Abstract
Camouflage object detection (COD) is designed to locate objects that are "seamlessly" embedded in the surrounding environment. Camouflaged object detection is a challenging task due to the high intrinsic similarities between objects and their backgrounds, as well as the low boundary contrast between them. To address this problem, this paper proposes a new ternary symmetric fusion network (TSFNet), which can detect camouflaged objects by fully fusing features of different levels and scales. Specifically, the network proposed in this paper mainly contains two key modules: the location-attention search (LAS) module and the ternary symmetric interaction fusion (TSIF) module. The location-attention search module makes full use of contextual information to position potential target objects from a global perspective while enhancing feature representation and guiding feature fusion. The ternary symmetric interaction fusion module consists of three branches: bilateral branches gather rich contextual information of multi-level features, and a middle branch provides fusion attention coefficients for the other two branches. The strategy can effectively achieve information fusion between low- and high-level features, and then achieve the refinement of edge details. Experimental results show that the method is an effective COD model and outperforms existing models. Compared with the existing model SINetV2, TSFNet significantly improves the performance by 3.5% weighted F-measure and 8.1% MAE on the COD10K. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
78. Editorial "special issue on artificial intelligence in practice – from theory to application".
- Author
-
Ali, Moonis, Friedrich, Gerhard, Pill, Ingo, and Wotawa, Franz
- Subjects
ARTIFICIAL intelligence ,DEEP learning ,JOB applications ,SPACE probes ,UNDERWATER cameras ,CASE-based reasoning - Abstract
The authors suggest to trade-off diagnosis quality versus runtime performance, exploiting learning for constraint ordering and where matrix factorization is used to estimate diagnoses. The final manuscript, "Applying matrix factorization to consistency-based direct diagnosis", authored by Seda Polat Erdeniz et al., deals with diagnosis as a basis for configuration problems when efficient solving is required in real-world applications. Screening all the accepted papers, the top 7 authors of the accepted papers were encouraged to submit their papers for this special issue, asking for substantial extensions to their submissions for creating manuscripts that are appropriate for being published in Applied Intelligence. [Extracted from the article]
- Published
- 2022
- Full Text
- View/download PDF
79. Prediction of stability coefficient of open-pit mine slope based on artificial intelligence deep learning algorithm.
- Author
-
Wang, Shuai, Zhang, Zongbao, and Wang, Chao
- Subjects
DEEP learning ,MACHINE learning ,ARTIFICIAL intelligence ,MINING engineering ,SLOPE stability ,MINE safety - Abstract
The mining of open pit mines is widespread in China, and there are many cases of landslide accidents. Therefore, the problem of slope stability is highlighted. The stability of the slope is a factor that directly affects the mining efficiency and the safety of the entire mining process. According to the statistics, there is a 15 percent chance of finding landslide risk in China's large-scale mines. And due to the expansion of the mining scale of the enterprise, the problem of slope stability has become increasingly obvious, which has become a major subject in the study of open-pit mine engineering. In order to better predict the slope stability coefficient, this study takes a mine in China as a case to deeply discuss the accuracy of different algorithms in the stability calculation, and then uses a deep learning algorithm to study the stability under rainfall conditions. The change of the coefficient and the change of the stability coefficient before and after the slope treatment are experimentally studied with the displacement of the monitoring point. The result shows that the safety coefficient calculated by the algorithm in this paper is about 7% lower than that of the traditional algorithm. In the slope stability analysis before treatment, the safety factor calculated by the algorithm in this paper is 1.086, and the algorithm in this paper is closer to reality. In the stability analysis of the slope after treatment, the safety factor calculated by the algorithm in this paper is 1.227, and the stability factor meets the requirements of the specification. It also shows that the deep learning algorithm effectively improves the efficiency of the slope stability factor prediction and improves security during project development. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
80. Explanation Paradigms Leveraging Analytic Intuition (ExPLAIn).
- Author
-
Jansen, Nils, Nolte, Gerrit, and Steffen, Bernhard
- Subjects
ARTIFICIAL neural networks ,INTUITION ,SOFTWARE development tools ,TECHNOLOGY transfer ,EXPLANATION - Abstract
In this paper, we present the envisioned style and scope of the new topic "Explanation Paradigms Leveraging Analytic Intuition" (ExPLAIn) with the International Journal on Software Tools for Technology Transfer (STTT). Intention behind this new topic is to (1) explicitly address all aspects and issues that arise when trying to, if possible, reveal and then confirm hidden properties of black-box systems, or (2) to enforce vital properties by embedding them into appropriate system contexts. Machine-learned systems, such as Deep Neural Networks, are particularly challenging black-box systems, and there is a wealth of formal methods for analysis and verification waiting to be adapted and applied. The selection of papers of this first Special Section of ExPLAIn, most of which were co-authored by editorial board members, is an illustrative example of the style and scope envisioned: In addition to methodological papers on verification, explanation, and their scalability, case studies, tool papers, literature reviews, and position papers are also welcome. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
81. A color attention mechanism based on YES color space for skin segmentation.
- Author
-
Ding, Shaobo, Liu, Zonghui, and Lei, Zhichun
- Abstract
Skin segmentation plays an important role in image processing and human–computer interaction tasks. However, it is a challenging task to accurately detect skin regions from various scenes with different illumination or color styles. In addition, in the field of video processing, reducing the computational load and improving the real-time performance of the algorithm has also become an important topic of skin segmentation. Existing deep semantic segmentation networks usually pay too much attention to the detection performance of the model and make the model structure tend to be complex, which brings heavy computational burden. To achieve the trade-off between detection performance and real-time performance of the skin segmentation algorithm, this paper proposes a lightweight skin segmentation network. Compared with existing semantic segmentation networks, this model adopts a simpler structure to improve the real-time performance. In addition, to improve the feature fitting ability of the network without slowing down its inference speed, this paper proposes a color attention mechanism, which locates skin regions in images based on the distribution features of skin colors on the E-R/G color plane generated from the YES color space, and guides the network to update parameters. Experimental results show that this method not only exhibits similar detection performance to existing semantic segmentation networks such as U-Net and DeepLab, but also the computation load of the model is 18.1% lower than Fast-SCNN. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
82. Real-time fire detection algorithms running on small embedded devices based on MobileNetV3 and YOLOv4.
- Author
-
Zheng, Hongtao, Duan, Junchen, Dong, Yu, and Liu, Yan
- Subjects
FIRE detectors ,DEEP learning ,FEATURE extraction ,ALGORITHMS ,NETWORK performance - Abstract
Copyright of Fire Ecology is the property of Springer Nature and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2023
- Full Text
- View/download PDF
83. A checklist to guide the planning, designing, implementation, and evaluation of learning analytics dashboards.
- Author
-
Kaliisa, Rogers, Jivet, Ioana, and Prinsloo, Paul
- Subjects
UNIVERSITIES & colleges ,DEEP learning - Abstract
Higher education institutions are moving to design and implement teacher-facing learning analytics (LA) dashboards with the hope that instructors can extract deep insights about student learning and make informed decisions to improve their teaching. While much attention has been paid to developing teacher-facing dashboards, less is known about how they are designed, implemented and evaluated. This paper presents a systematic literature review of existing studies reporting on teacher-facing LA dashboards. Out of the 1968 articles retrieved from several databases, 50 articles were included in the final analysis. Guided by several frameworks, articles were coded based on the following dimensions: purpose, theoretical grounding, stakeholder involvement, ethics and privacy, design, implementation, and evaluation criteria. The findings show that most dashboards are designed to increase teachers' awareness but with limited actionable insights to allow intervention. Moreover, while teachers are involved in the design process, this is mainly at the exploratory/problem definition stage, with little input beyond this stage. Most dashboards were prescriptive, less customisable, and implicit about the theoretical constructs behind their designs. In addition, dashboards are deployed at prototype and pilot stages, and the evaluation is dominated by self-reports and users' reactions with limited focus on changes to teaching and learning. Besides, only one study considered privacy as a design requirement. Based on the findings of the study and synthesis of existing literature, we propose a four-dimensional checklist for planning, designing, implementing and evaluating LA dashboards. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
84. Union-net: lightweight deep neural network model suitable for small data sets.
- Author
-
Zhou, Jingyi, He, Qingfang, Cheng, Guang, and Lin, Zhiying
- Subjects
DEEP learning ,BIG data - Abstract
Traditional deep learning models prefer large data sets, and in reality small data sets are easier to obtain. It is more practical to build models suitable for small data sets. Based on CNN, this paper proposes the concept of union convolution to build a deep learning model Union-net that is suitable for small data sets. The Union-net has small model size and superior performance. In this paper, the model is tested based on multiple commonly used data sets. The experimental results show that Union-net outperforms most models when dealing with small datasets, and Union-net outperforms other models when dealing with complex classification tasks or dealing with few-shot datasets. The codes for this paper have been uploaded to https://github.com/yeaso/union-net. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
85. Comparative performance of fully-automated and semi-automated artificial intelligence methods for the detection of clinically significant prostate cancer on MRI: a systematic review.
- Author
-
Sushentsev, Nikita, Moreira Da Silva, Nadia, Yeung, Michael, Barrett, Tristan, Sala, Evis, Roberts, Michael, and Rundo, Leonardo
- Subjects
ARTIFICIAL intelligence ,PROSTATE cancer ,MACHINE learning ,MAGNETIC resonance imaging ,MEDICAL screening - Abstract
Objectives: We systematically reviewed the current literature evaluating the ability of fully-automated deep learning (DL) and semi-automated traditional machine learning (TML) MRI-based artificial intelligence (AI) methods to differentiate clinically significant prostate cancer (csPCa) from indolent PCa (iPCa) and benign conditions. Methods: We performed a computerised bibliographic search of studies indexed in MEDLINE/PubMed, arXiv, medRxiv, and bioRxiv between 1 January 2016 and 31 July 2021. Two reviewers performed the title/abstract and full-text screening. The remaining papers were screened by four reviewers using the Checklist for Artificial Intelligence in Medical Imaging (CLAIM) for DL studies and Radiomics Quality Score (RQS) for TML studies. Papers that fulfilled the pre-defined screening requirements underwent full CLAIM/RQS evaluation alongside the risk of bias assessment using QUADAS-2, both conducted by the same four reviewers. Standard measures of discrimination were extracted for the developed predictive models. Results: 17/28 papers (five DL and twelve TML) passed the quality screening and were subject to a full CLAIM/RQS/QUADAS-2 assessment, which revealed a substantial study heterogeneity that precluded us from performing quantitative analysis as part of this review. The mean RQS of TML papers was 11/36, and a total of five papers had a high risk of bias. AUCs of DL and TML papers with low risk of bias ranged between 0.80–0.89 and 0.75–0.88, respectively. Conclusion: We observed comparable performance of the two classes of AI methods and identified a number of common methodological limitations and biases that future studies will need to address to ensure the generalisability of the developed models. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
86. Biometric-based unimodal and multimodal person identification with CNN using optimal filter set.
- Author
-
Sarker, Goutam and Ghosh, Swagata
- Abstract
The convolutional neural network (CNN) has brought about a drastic change in the field of image processing and pattern recognition. The filters of CNN model correspond to the activation maps that extract features from the input images. Thus, the number of filters and filter size are of significant importance to learning and recognition accuracy of CNN model-based systems such as the biometric-based person authentication system. The present paper proposes to analyze the impact of varying the number of filters of CNN models on the accuracy of the biometric-based single classifiers using human face, fingerprint and iris for person identification, and also biometric-based super-classification using both bagging- and programming-based boosting methods. The present paper gives an insight to the optimal set of filters in CNN model that gives the maximum overall accuracy of the classifier system. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
87. Efficient evolutionary neural architecture search based on hybrid search space.
- Author
-
Gong, Tao, Ma, Yongjie, Xu, Yang, and Song, Changwei
- Abstract
Manually designed convolutional neural networks have demonstrated excellent performance in various domains, but designing neural networks suitable for specific tasks poses significant challenges, and the emergence of Neural Structure Search (NAS) provides a new solution to this problem. However, existing algorithms either focus solely on network lightweight, resulting in subpar network performance, or excessively emphasize performance, leading to substantial network redundancy. With consideration for both network parameters and performance, this paper designs a hybrid search space based on residual modules and RepVGG modules using genetic algorithm, and stacks them together to form a more efficient network. To achieve this, we propose an efficient variable-length encoding strategy, utilizing units as the fundamental encoding space to encode variable-length individuals; we design evolutionary operations encompassing single-point crossover and three types of mutation operators to ensure population diversity; during training, a random forest-based performance predictor is employed to significantly shorten the network search time. To demonstrate the effectiveness of the proposed algorithm, we introduce the concept of transfer learning, which involves decoding the globally optimal solution, fine-tuning it, and then transferring it to three categories of real-world application datasets. Through comparisons with various algorithms, our approach consistently achieved leading performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
88. Production quality prediction of cross-specification products using dynamic deep transfer learning network.
- Author
-
Wang, Pei, Wang, Tao, Yang, Sheng, Cheng, Han, Huang, Pengde, and Zhang, Qianle
- Subjects
TECHNICAL specifications ,DATA distribution ,PREDICTION models ,FORECASTING ,NEW product development ,DEEP learning ,SUPERVISED learning - Abstract
In the process of industrial production, products with different specifications (i.e., the difference in geometry, process conditions, and machine conditions, etc.) have different quality data distributions, which lead to a decrease in the accuracy of traditional data-driven quality prediction models that require the same quality data distribution. At the same time, due to economic cost factors, obtaining a large amount of accumulated data for different specifications is difficult, and the re-modeling data accumulation of multiple cross-specifications is insufficient. In order to solve the quality prediction problem of production with different data distributions and poor data accumulation, we use the deep transfer learning (DTL) method with unsupervised dynamic domain adaptation (DDA) to transfer the domain invariant features learned from labeled specification products (source domain) to other unlabeled new specification products (target domain). In order to improve the success rate of cross-domain quality prediction, the Wasserstein distance adapter is designed to match appropriate source domain samples and target domain samples to build multiple transfer tasks that are suitable for transfer. At the same time, the dynamic distribution adaptation and dynamic adversarial adaptation are combined to extract the domain invariant features to improve the adaptability of the prediction model for products with new specifications (e.g., size difference) and unlabeled and limited quality data. Finally, a comprehensive experiment is carried out using the actual production data of products with different specifications. The experimental results show that compared with the traditional non-transfer deep learning methods, the MAE, RMSE, and R
2 of the proposed DTL method are improved by 18.26%, 16.66%, and 22.48% respectively. Compared with other transfer methods, the MAE, RMSE, and R2 of the DTL proposed in this paper are improved by 10.45%, 10.96%, and 9.72% respectively. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
89. Deep learning-based methods for natural hazard named entity recognition.
- Author
-
Sun, Junlin, Liu, Yanrong, Cui, Jing, and He, Handong
- Subjects
DEEP learning ,HAZARD mitigation ,HAZARDS ,FEATURE extraction - Abstract
Natural hazard named entity recognition is a technique used to recognize natural hazard entities from a large number of texts. The method of natural hazard named entity recognition can facilitate acquisition of natural hazards information and provide reference for natural hazard mitigation. The method of named entity recognition has many challenges, such as fast change, multiple types and various forms of named entities. This can introduce difficulties in research of natural hazard named entity recognition. To address the above problem, this paper constructed a natural disaster annotated corpus for training and evaluation model, and selected and compared several deep learning methods based on word vector features. A deep learning method for natural hazard named entity recognition can automatically mine text features and reduce the dependence on manual rules. This paper compares and analyzes the deep learning models from three aspects: pretraining, feature extraction and decoding. A natural hazard named entity recognition method based on deep learning is proposed, namely XLNet-BiLSTM-CRF model. Finally, the research hotspots of natural hazards papers in the past 10 years were obtained through this model. After training, the precision of the XLNet-BilSTM-CRF model is 92.80%, the recall rate is 91.74%, and the F1-score is 92.27%. The results show that this method, which is superior to other methods, can effectively recognize natural hazard named entities. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
90. Crowd density estimation based on multi scale features fusion network with reverse attention mechanism.
- Author
-
Li, Yong-Chao, Jia, Rui-Sheng, Hu, Ying-Xiang, Han, Dong-Nuo, and Sun, Hong-Mei
- Subjects
CONVOLUTIONAL neural networks ,FEATURE extraction ,DENSITY ,CROWDS ,DEEP learning ,COMPOSITE columns ,MACHINE learning - Abstract
Deep learning has made substantial progress in crowd counting, but in practical applications, due to interference factors such as perspective distortion and complex background, the existing methods still have large errors in counting. In response to the above problems, this paper designs a multi-scale feature fusion network (IA-MFFCN) based on the reverse attention mechanism, which maps the image to the crowd density map for counting. The network consists of three parts: feature extraction module, inverse attention module, and back-end module. First, to overcome the problem of perspective distortion, deeper single-column CNNs was designed as a feature extraction module to extract multi-scale feature information and merge them; second, to avoid interference of complex backgrounds, the inverse attention module was designed, through the multi-scale inverse attention mechanism, reducing the influence of noise on counting accuracy. Finally, to generate a high-quality crowd density map, dilation convolution was introduced. Simultaneously, to enhance the sensitivity of the network to crowd counting, a comprehensive loss function based on Euclidean loss and predicted population loss is designed to improve training accuracy, to produce a more accurate density value. Experiments show that compared with the comparison algorithm, the algorithm in this paper has a significant reduction in the mean absolute error (MAE) and mean square error (MSE) on the ShanghaiTech dataset, UCF_CC_50 dataset and WorldExpo'10 dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
91. Learning symmetry-aware atom mapping in chemical reactions through deep graph matching.
- Author
-
Astero, Maryam and Rousu, Juho
- Subjects
CHEMICAL reactions ,GRAPH neural networks ,REPRESENTATIONS of graphs ,MOLECULAR graphs ,ATOMS ,GRAPH algorithms - Abstract
Accurate atom mapping, which establishes correspondences between atoms in reactants and products, is a crucial step in analyzing chemical reactions. In this paper, we present a novel end-to-end approach that formulates the atom mapping problem as a deep graph matching task. Our proposed model, AMNet (Atom Matching Network), utilizes molecular graph representations and employs various atom and bond features using graph neural networks to capture the intricate structural characteristics of molecules, ensuring precise atom correspondence predictions. Notably, AMNet incorporates the consideration of molecule symmetry, enhancing accuracy while simultaneously reducing computational complexity. The integration of the Weisfeiler-Lehman isomorphism test for symmetry identification refines the model's predictions. Furthermore, our model maps the entire atom set in a chemical reaction, offering a comprehensive approach beyond focusing solely on the main molecules in reactions. We evaluated AMNet's performance on a subset of USPTO reaction datasets, addressing various tasks, including assessing the impact of molecular symmetry identification, understanding the influence of feature selection on AMNet performance, and comparing its performance with the state-of-the-art method. The result reveals an average accuracy of 97.3% on mapped atoms, with 99.7% of reactions correctly mapped when the correct mapped atom is within the top 10 predicted atoms. Scientific contribution The paper introduces a novel end-to-end deep graph matching model for atom mapping, utilizing molecular graph representations to capture structural characteristics effectively. It enhances accuracy by integrating symmetry detection through the Weisfeiler-Lehman test, reducing the number of possible mappings and improving efficiency. Unlike previous methods, it maps the entire reaction, not just main components, providing a comprehensive view. Additionally, by integrating efficient graph matching techniques, it reduces computational complexity, making atom mapping more feasible. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
92. Integrated encoder-decoder-based wide and deep convolution neural networks strategy for electricity theft arbitration.
- Author
-
Kumawat, Manoj, Onaolapo, Adeniyi, Sharma, Gulshan, Barutcu, Ibrahim Cagri, Adefarati, Temitope, and Bansal, Ramesh
- Subjects
CONVOLUTIONAL neural networks ,MACHINE learning ,THEFT ,DEEP learning ,ENTORHINAL cortex ,ELECTRICITY - Abstract
Integrating energy systems with information systems in smart grids offers a promising avenue for combating electricity theft by leveraging real-time data insights. Suspicious activity indicative of theft can be identified through anomalous consumption patterns observed in smart networks. However, a smart model is required for capturing and analysing the data intelligently to accurately detect electricity theft. In the paper, electricity theft has been detected using an encoder-decoder-based classifier that integrates two models of convolutional neural networks (CNN). The aim is to scan the strength of the data and built a smart model that analysed the connections in complex data and determine the pattern of theft. The model comprises three compartments: the auto-encoder, the wide convolutional neural network (1-D CNN model), and the deep convolutional neural network (2-D CNN model). The auto-encoder has been trained on the complex and in-depth linkage between the theft data and the normal data as it removes noise and unnecessary information. The 1-D CNN model gathers relevant connections and general features, while the 2-D CNN model determines the rate at which energy theft occurs and differentiates between the energy-stealing consumers and normal consumers. The efficacy of the approach is underscored by its superiority over traditional deep learning and machine learning techniques. This paper elucidates the distinct advantages and applications of the proposed model in combating electricity theft within smart grid environments. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
93. Relationship constraint deep metric learning.
- Author
-
Zhang, Yanbing, Xiao, Ting, Wang, Zhe, Wang, Xinru, Feng, Wenyi, Fu, Zhiling, and Yang, Hai
- Subjects
DEEP learning ,TIME complexity ,PROBLEM solving ,CLASSIFICATION algorithms - Abstract
Deep metric learning (DML) models aim to learn semantically meaningful representations in which similar samples are pulled together and dissimilar samples are pushed apart. However, the classification effect is limited due to the high time complexity of previous models and their poor performance in extracting data relationships. This paper presents a novel relationship constraint deep metric learning (RCDML) approach, including proxy relationship constraint (PRC) and sample relationship constraint (SRC) for inter-class separability and intra-class compactness, to solve the above problems and improve the classification effect. The PRC combines the proxy-to-proxy relationship loss term with the proxy-to-sample relationship loss function to maximize the proxy features, hence enhancing inter-class separability by decreasing proxy similarity. Additionally, the SRC combines the sample-to-sample relationship loss term with the proxy-to-sample relationship loss function to maximize the sample features, which promotes intra-class compactness by increasing the similarity between the most different samples of the same class. Unlike existing proxy-based and pair-based methods, the relationship constraint framework uses a diverse range of proxy and sample data relationships. In addition, the proxy correction (PC) module is used to optimize the proxy. Extensive tests conducted on the widely popular CUB-200-2011, CARS-196, and SOP datasets show that the framework is effective and attains state-of-the-art performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
94. Forward layer-wise learning of convolutional neural networks through separation index maximizing.
- Author
-
Karimi, Ali, Kalhor, Ahmad, and Sadeghi Tabrizi, Melika
- Subjects
CONVOLUTIONAL neural networks ,MACHINE learning ,IMAGE recognition (Computer vision) ,TIME complexity ,DEEP learning ,CLASSIFICATION algorithms - Abstract
This paper proposes a forward layer-wise learning algorithm for CNNs in classification problems. The algorithm utilizes the Separation Index (SI) as a supervised complexity measure to evaluate and train each layer in a forward manner. The proposed method explains that gradually increasing the SI through layers reduces the input data's uncertainties and disturbances, achieving a better feature space representation. Hence, by approximating the SI with a variant of local triplet loss at each layer, a gradient-based learning algorithm is suggested to maximize it. Inspired by the NGRAD (Neural Gradient Representation by Activity Differences) hypothesis, the proposed algorithm operates in a forward manner without explicit error information from the last layer. The algorithm's performance is evaluated on image classification tasks using VGG16, VGG19, AlexNet, and LeNet architectures with CIFAR-10, CIFAR-100, Raabin-WBC, and Fashion-MNIST datasets. Additionally, the experiments are applied to text classification tasks using the DBPedia and AG's News datasets. The results demonstrate that the proposed layer-wise learning algorithm outperforms state-of-the-art methods in accuracy and time complexity. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
95. Feature extraction of multimodal medical image fusion using novel deep learning and contrast enhancement method.
- Author
-
Bhutto, Jameel Ahmed, Guosong, Jiang, Rahman, Ziaur, Ishfaq, Muhammad, Sun, Zhengzheng, and Soomro, Toufique Ahmed
- Subjects
IMAGE fusion ,FEATURE extraction ,COMPUTER-assisted image analysis (Medicine) ,CONVOLUTIONAL neural networks ,DIAGNOSTIC imaging ,DEEP learning ,MULTIMODAL user interfaces - Abstract
The fusion of multimodal medical images has garnered painstaking attention for clinical diagnosis and surgical planning. Although various scholars have designed numerous fusion methods, the challenges of extracting substantial features without introducing noise and non-uniform contrast hindered the overall quality of fused photos. This paper presents a multimodal medical image fusion (MMIF) using a novel deep convolutional neural network (D-CNN) along with preprocessing schemes to circumvent the mentioned issues. A non-linear average median filtering (NL-AMF) and multiscale improved top-hat (MI-TH) approach are utilized at the preprocessing stage to remove noise and improve the contrast of images. The non-linear anisotropic diffusion (NL-AD) scheme is employed to split the photos into base and detailed parts. The fusion of base parts is accomplished by a dimension reduction method to retain the energy information. In contrast, the detailed parts are fused by novel D-CNN to preserve the enriched detailed features effectively. The simulation results demonstrate that the proposed method produces better brightness contrast and more image details than existing methods by acquiring 0.7649 to 0.8986, 0.3520 to 0.4783, 0.7639 to 0.9056, 68.8932 to 81.0487 gain for quality transfer ratio from source photo to a generated photo ( Q G AB ), feature mutual information (FMI), structural similarity index (SSIM), and average pixel intensity (API) respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
96. Universal approximation property of stochastic configuration networks for time series.
- Author
-
Zhang, Jin-Xi, Zhao, Hangyi, and Zhang, Xuefeng
- Subjects
RECURRENT neural networks ,TIME series analysis ,STOCHASTIC approximation ,MACHINE learning ,SEQUENTIAL learning ,ELECTRONIC data processing - Abstract
For the purpose of processing sequential data, such as time series, and addressing the challenge of manually tuning the architecture of traditional recurrent neural networks (RNNs), this paper introduces a novel approach-the Recurrent Stochastic Configuration Network (RSCN). This network is constructed based on the random incremental algorithm of stochastic configuration networks. Leveraging the foundational structure of recurrent neural networks, our learning model commences with a modest-scale recurrent neural network featuring a single hidden layer and a solitary hidden node. Subsequently, the node parameters of the hidden layer undergo incremental augmentation through a random configuration process, with corresponding weights assigned structurally. This iterative expansion continues until the network satisfies predefined termination criteria. Noteworthy is the adaptability of this algorithm to handle time series data, exhibiting superior performance compared to traditional recurrent neural networks with similar architectures. The experimental results presented in this paper underscore the efficacy of the proposed RSCN for sequence data processing, showcasing its advantages over conventional recurrent neural networks in the context of the performed experiments. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
97. Comparing deep learning and pathologist quantification of cell-level PD-L1 expression in non-small cell lung cancer whole-slide images.
- Author
-
van Eekelen, Leander, Spronck, Joey, Looijen-Salamon, Monika, Vos, Shoko, Munari, Enrico, Girolami, Ilaria, Eccher, Albino, Acs, Balazs, Boyaci, Ceren, de Souza, Gabriel Silva, Demirel-Andishmand, Muradije, Meesters, Luca Dulce, Zegers, Daan, van der Woude, Lieke, Theelen, Willemijn, van den Heuvel, Michel, Grünberg, Katrien, van Ginneken, Bram, van der Laak, Jeroen, and Ciompi, Francesco
- Subjects
DEEP learning ,NON-small-cell lung carcinoma ,PROGRAMMED death-ligand 1 ,MACHINE learning ,COMPUTER vision ,PATHOLOGISTS - Abstract
Programmed death-ligand 1 (PD-L1) expression is currently used in the clinic to assess eligibility for immune-checkpoint inhibitors via the tumor proportion score (TPS), but its efficacy is limited by high interobserver variability. Multiple papers have presented systems for the automatic quantification of TPS, but none report on the task of determining cell-level PD-L1 expression and often reserve their evaluation to a single PD-L1 monoclonal antibody or clinical center. In this paper, we report on a deep learning algorithm for detecting PD-L1 negative and positive tumor cells at a cellular level and evaluate it on a cell-level reference standard established by six readers on a multi-centric, multi PD-L1 assay dataset. This reference standard also provides for the first time a benchmark for computer vision algorithms. In addition, in line with other papers, we also evaluate our algorithm at slide-level by measuring the agreement between the algorithm and six pathologists on TPS quantification. We find a moderately low interobserver agreement at cell-level level (mean reader-reader F1 score = 0.68) which our algorithm sits slightly under (mean reader-AI F1 score = 0.55), especially for cases from the clinical center not included in the training set. Despite this, we find good AI-pathologist agreement on quantifying TPS compared to the interobserver agreement (mean reader-reader Cohen's kappa = 0.54, 95% CI 0.26–0.81, mean reader-AI kappa = 0.49, 95% CI 0.27—0.72). In conclusion, our deep learning algorithm demonstrates promise in detecting PD-L1 expression at a cellular level and exhibits favorable agreement with pathologists in quantifying the tumor proportion score (TPS). We publicly release our models for use via the Grand-Challenge platform. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
98. A new method for deep learning detection of defects in X-ray images of pressure vessel welds.
- Author
-
Wang, Xue, He, Feng, and Huang, Xu
- Subjects
PRESSURE vessels ,X-ray imaging ,X-ray detection ,WELDED joints ,VALUE engineering ,DEEP learning ,IMAGE segmentation - Abstract
Given that defect detection in weld X-ray images is a critical aspect of pressure vessel manufacturing and inspection, accurate differentiation of the type, distribution, number, and area of defects in the images serves as the foundation for judging weld quality, and the segmentation method of defects in digital X-ray images is the core technology for differentiating defects. Based on the publicly available weld seam dataset GDX-ray, this paper proposes a complete technique for fault segmentation in X-ray pictures of pressure vessel welds. The key works are as follows: (1) To address the problem of a lack of defect samples and imbalanced distribution inside GDX-ray, a DA-DCGAN based on a two-channel attention mechanism is devised to increase sample data. (2) A convolutional block attention mechanism is incorporated into the coding layer to boost the accuracy of small-scale defect identification. The proposed MAU-Net defect semantic segmentation network uses multi-scale even convolution to enhance large-scale features. The proposed method can mask electrostatic interference and non-defect-class parts in the actual weld X-ray images, achieve an average segmentation accuracy of 84.75% for the GDX-ray dataset, segment and accurately rate the valid defects with a correct rating rate of 95%, and thus realize practical value in engineering. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
99. A robot grasping detection network based on flexible selection of multi-modal feature fusion structure.
- Author
-
Wang, Yuhan, Guo, Zhibo, Chen, Yu, Guo, Chaiqi, Xia, Meizhen, and Qi, Tingyue
- Subjects
ROBOT hands ,ROBOTS ,DEEP learning ,ROBOTICS ,PYRAMIDS - Abstract
In unstructured scenarios, objects usually have unique shapes, poses, and other uncertainties, which put forward higher requirements for the robot's planar grasping detection ability. Most previous methods use single-modal data or simply fused multi-modal data to predict gripper configurations. Single-modal data is not conducive to comprehensively describe the diversity of objects, and the simple fusion method may also ignore the dependencies between multi-modal data. Based on the above considerations, we propose a Multi-modal Dynamic Cooperative Fusion Network (MDCNet), in which a Multilevel Semantic Guided Fusion Module (MSG) is designed, through which enhanced semantic guidance vectors are used to suppress the undesired influence factors produced by different fusion structures. In addition, we also design a general Enhanced Feature Pyramid Nets Structure (EFPN) to learn the dependencies between fine-grained features and coarse-grained features and improve the robustness of the encoder in unstructured scenarios. The results show that the proposed method has an accuracy rate of 98.9% on the Jacquard dataset and 99.6% on the Cornell dataset. In over 2000 robotic grasp trials, our structure achieves a grasp success rate of 98.8% in single-object scenarios and 93.5% in cluttered scenarios. The proposed method in this paper is superior to previous grasp detection methods in both speed and accuracy, and has strong real-time performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
100. BiLSTM-TANet: an adaptive diverse scenes model with context embeddings for few-shot learning.
- Author
-
Zhang, He, Liu, Han, Liang, Lili, Ma, Wenlu, and Liu, Ding
- Subjects
DEEP learning ,COMPUTER vision ,FEATURE extraction ,EUCLIDEAN distance - Abstract
Few-shot learning is a critical task in computer vision processing that helps reduce deep learning's reliance on large datasets. This paper aims to establish a few-shot learning network that is adaptive to diverse scenes. A novel approach referred to as task-adapted network with bi-directional long short-term memory network (BiLSTM-TANet) is proposed in this paper. BiLSTM-TANet is an end-to-end approach based on deep metric learning and designed to use the information from finite samples as much as possible. It fuses the context embeddings and structure information of the images and adaptively adjusts the features several times during the feature extraction of task to achieve task-specific embedding and quickly adapt to different distributed tasks, improves the feature extraction performance, and strikes a balance between model stability and generality. The model employs Euclidean distance as the classifier to reduce the number of model parameters and enhance the classification performance. Experiments conducted on miniImageNet, TieredImageNet, CUB200_2011 and CIFAR-FS datasets demonstrate the performance of the proposed BiLSTM-TANet. Furthermore, the effects of different few-shot learning parameters on the model's performance are explored, providing a helpful reference for the future study of few-shot learning. Finally, a series of ablation studies are performed to analyze the performance of BiLSTM-TANet. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.