1,703 results
Search Results
2. Citation recommendation using modified HITS algorithm.
- Author
-
Kammari, Monachary and Bhavani, S. Durga
- Subjects
DEEP learning ,ALGORITHMS ,COMPUTER performance ,WEBSITES ,MACHINE learning - Abstract
Over the years the number of research publications per year is growing exponentially. Finding research papers of quality from the massive literature of relevant articles is a challenging and time-consuming task. The approaches in the latest literature address citation recommendation by utilizing large bibliographic information and use machine learning and deep learning methods for the task. These techniques clearly require a large amount of training data as well as machines with high processing power. To overcome these issues, we propose a novel method by modifying the popular hyperlink induced topic search (HITS), a web page ranking algorithm, as citation recommendation using hyperlink induced topic search (CR-HITS) that works on a directed and weighted heterogeneous bibliographic network containing diverse types of nodes and edges. We define effective scoring schemes for nodes and edges based on basic bibliographic information like citations of papers, number of publications of an author, etc. Given a few seed papers, the citation recommendation algorithm CR-HITS is run on small neighborhoods of the seed papers and hence the time taken by the execution is very small to yield the final recommendations. To the best of our knowledge, HITS has been used for the first time for the citation recommendation problem. We perform extensive experimentation on DBLP (version-11) and ACM (version-9) datasets and compare the results with many baseline methods in terms of MAP, MRR, and recall@N measures. The performance of the proposed algorithms is superior with respect to the MAP metric and matches the second best for the other two metrics. Since the top two algorithms use deep learning methods and use much larger bibliographic information including abstracts of the papers, we claim that our approach utilizes very low resources, yet yields recommendations that are very close to the top recommendations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Reproducibility of Deep Learning Algorithms Developed for Medical Imaging Analysis: A Systematic Review.
- Author
-
Moassefi, Mana, Rouzrokh, Pouria, Conte, Gian Marco, Vahdati, Sanaz, Fu, Tianyuan, Tahmasebi, Aylin, Younis, Mira, Farahani, Keyvan, Gentili, Amilcare, Kline, Timothy, Kitamura, Felipe C., Huo, Yuankai, Kuanar, Shiba, Younis, Khaled, Erickson, Bradley J., and Faghani, Shahriar
- Subjects
DEEP learning ,RESEARCH evaluation ,SYSTEMATIC reviews ,ARTIFICIAL intelligence ,DIAGNOSTIC imaging ,DESCRIPTIVE statistics ,ALGORITHMS ,WORLD Wide Web - Abstract
Since 2000, there have been more than 8000 publications on radiology artificial intelligence (AI). AI breakthroughs allow complex tasks to be automated and even performed beyond human capabilities. However, the lack of details on the methods and algorithm code undercuts its scientific value. Many science subfields have recently faced a reproducibility crisis, eroding trust in processes and results, and influencing the rise in retractions of scientific papers. For the same reasons, conducting research in deep learning (DL) also requires reproducibility. Although several valuable manuscript checklists for AI in medical imaging exist, they are not focused specifically on reproducibility. In this study, we conducted a systematic review of recently published papers in the field of DL to evaluate if the description of their methodology could allow the reproducibility of their findings. We focused on the Journal of Digital Imaging (JDI), a specialized journal that publishes papers on AI and medical imaging. We used the keyword "Deep Learning" and collected the articles published between January 2020 and January 2022. We screened all the articles and included the ones which reported the development of a DL tool in medical imaging. We extracted the reported details about the dataset, data handling steps, data splitting, model details, and performance metrics of each included article. We found 148 articles. Eighty were included after screening for articles that reported developing a DL model for medical image analysis. Five studies have made their code publicly available, and 35 studies have utilized publicly available datasets. We provided figures to show the ratio and absolute count of reported items from included studies. According to our cross-sectional study, in JDI publications on DL in medical imaging, authors infrequently report the key elements of their study to make it reproducible. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
4. A term extraction algorithm based on machine learning and comprehensive feature strategy.
- Author
-
Gong, Xiuliang, Cheng, Bo, Hu, Xiaomei, and Bo, Wen
- Subjects
MACHINE learning ,NATURAL language processing ,ALGORITHMS ,RANDOM fields ,ONTOLOGIES (Information retrieval) ,DATABASES ,MACHINE translating - Abstract
Manual term extraction is similar to literal meaning: A translator browses text, classifies words, and prepares for translation. Terminology, as a centralized carrier of expertise, creation, popularization, and disappearance, dynamically reflects the development and evolution of an industry. The automatic extraction of terminology is a key technology for creating a professional terminology database, and it is also a key topic in the field of natural language processing. The purpose of this paper is to study how to analyse a term extraction algorithm based on machine learning and a comprehensive feature strategy. Focusing on the problems of poor generality and single statistical features of current term extraction algorithms, this paper proposes an improved domain ontology term extraction algorithm based on a comprehensive feature strategy. Moreover, automatic term extraction experiments based on a word-based maximum entropy model and a conditional random field model based on machine learning are conducted in this paper. Its word-based conditional random field model outperforms the maximum entropy model. The experimental results show that the algorithm based on the comprehensive feature strategy improves the accuracy by 8.6% compared with the TF-IDF algorithm and the C-value term extraction algorithm. This algorithm can be used to effectively extract the terms in a text and has good generality. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Reviewing Machine Learning and Image Processing Based Decision-Making Systems for Breast Cancer Imaging.
- Author
-
Zerouaoui, Hasnae and Idri, Ali
- Subjects
BREAST tumor diagnosis ,ALGORITHMS ,MAMMOGRAMS ,BREAST tumors ,DECISION support systems ,DECISION trees ,DIAGNOSTIC imaging ,DIGITAL image processing ,MACHINE learning ,MAGNETIC resonance imaging ,MEDLINE ,ARTIFICIAL neural networks ,ONLINE information services ,RESEARCH funding ,SYSTEMATIC reviews ,RESEARCH bias ,SUPPORT vector machines ,DESCRIPTIVE statistics ,COMPUTER-aided diagnosis ,DEEP learning - Abstract
Breast cancer (BC) is the leading cause of death among women worldwide. It affects in general women older than 40 years old. Medical images analysis is one of the most promising research areas since it provides facilities for diagnosis and decision-making of several diseases such as BC. This paper conducts a Structured Literature Review (SLR) of the use of Machine Learning (ML) and Image Processing (IP) techniques to deal with BC imaging. A set of 530 papers published between 2000 and August 2019 were selected and analyzed according to ten criteria: year and publication channel, empirical type, research type, medical task, machine learning techniques, datasets used, validation methods, performance measures and image processing techniques which include image pre-processing, segmentation, feature extraction and feature selection. Results showed that diagnosis was the most used medical task and that Deep Learning techniques (DL) were largely used to perform classification. Furthermore, we found out that classification was the most ML objective investigated followed by prediction and clustering. Most of the selected studies used Mammograms as imaging modalities rather than Ultrasound or Magnetic Resonance Imaging with the use of public or private datasets with MIAS as the most frequently investigated public dataset. As for image processing techniques, the majority of the selected studies pre-process their input images by reducing the noise and normalizing the colors, and some of them use segmentation to extract the region of interest with the thresholding method. For feature extraction, we note that researchers extracted the relevant features using classical feature extraction techniques (e.g. Texture features, Shape features, etc.) or DL techniques (e. g. VGG16, VGG19, ResNet, etc.), and finally few papers used feature selection techniques in particular the filter methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
6. What do algorithms explain? The issue of the goals and capabilities of Explainable Artificial Intelligence (XAI).
- Author
-
Renftle, Moritz, Trittenbach, Holger, Poznic, Michael, and Heil, Reinhard
- Subjects
ARTIFICIAL intelligence ,MACHINE learning ,ALGORITHMS - Abstract
The increasing ubiquity of machine learning (ML) motivates research on algorithms to "explain" models and their predictions—so-called Explainable Artificial Intelligence (XAI). Despite many publications and discussions, the goals and capabilities of such algorithms are far from being well understood. We argue that this is because of a problematic reasoning scheme in the literature: Such algorithms are said to complement machine learning models with desired capabilities, such as interpretability or explainability. These capabilities are in turn assumed to contribute to a goal, such as trust in a system. But most capabilities lack precise definitions and their relationship to such goals is far from obvious. The result is a reasoning scheme that obfuscates research results and leaves an important question unanswered: What can one expect from XAI algorithms? In this paper, we clarify the modest capabilities of these algorithms from a concrete perspective: that of their users. We show that current algorithms can only answer user questions that can be traced back to the question: "How can one represent an ML model as a simple function that uses interpreted attributes?". Answering this core question can be trivial, difficult or even impossible, depending on the application. The result of the paper is the identification of two key challenges for XAI research: the approximation and the translation of ML models. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Machine Learning Models to Predict Readmission Risk of Patients with Schizophrenia in a Spanish Region.
- Author
-
Góngora Alonso, Susel, Herrera Montano, Isabel, Ayala, Juan Luis Martín, Rodrigues, Joel J. P. C., Franco-Martín, Manuel, and de la Torre Díez, Isabel
- Subjects
MACHINE learning ,MENTAL health services ,PATIENT readmissions ,PEOPLE with schizophrenia ,PUBLIC hospitals - Abstract
Currently, high hospital readmission rates have become a problem for mental health services, because it is directly associated with the quality of patient care. The development of predictive models with machine learning algorithms allows the assessment of readmission risk in hospitals. The main objective of this paper is to predict the readmission risk of patients with schizophrenia in a region of Spain, using machine learning algorithms. In this study, we used a dataset with 6089 electronic admission records corresponding to 3065 patients with schizophrenia disorders. Data were collected in the period 2005–2015 from acute units of 11 public hospitals in a Spain region. The Random Forest classifier obtained the best results in predicting the readmission risk, in the metrics accuracy = 0.817, recall = 0.887, F1-score = 0.877, and AUC = 0.879. This paper shows the algorithm with highest accuracy value and determines the factors associated with readmission risk of patients with schizophrenia in this population. It also shows that the development of predictive models with a machine learning approach can help improve patient care quality and develop preventive treatments. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. An effective video inpainting technique using morphological Haar wavelet transform with krill herd based criminisi algorithm.
- Author
-
Srinivasan, M. Nuthal, Chinnadurai, M., Senthilkumar, S., and Dinesh, E.
- Subjects
WAVELET transforms ,MACHINE learning ,INPAINTING ,ANIMAL herds ,ALGORITHMS ,SIGNAL-to-noise ratio - Abstract
In recent times, video inpainting techniques have intended to fill the missing areas or gaps in a video by utilizing known pixels. The variety in brightness or difference of the patches causes the state-of-the-art video inpainting techniques to exhibit high computation complexity and create seams in the target areas. To resolve these issues, this paper introduces a novel video inpainting technique that employs the Morphological Haar Wavelet Transform combined with the Krill Herd based Criminisi algorithm (MHWT-KHCA) to address the challenges of high computational demand and visible seam artifacts in current inpainting practices. The proposed MHWT-KHCA algorithm strategically reduces computation times and enhances the seamlessness of the inpainting process in videos. Through a series of experiments, the technique is validated against standard metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM), where it demonstrates superior performance compared to existing methods. Additionally, the paper outlines potential real-world applications ranging from video restoration to real-time surveillance enhancement, highlighting the technique's versatility and effectiveness. Future research directions include optimizing the algorithm for diverse video formats and integrating machine learning models to advance its capabilities further. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Selected and Extended Papers from TACAS 2018: Preface.
- Author
-
Beyer, Dirk and Huisman, Marieke
- Subjects
SOFTWARE development tools ,DATA structures ,ALGORITHMS ,MACHINE learning ,DYNAMIC programming - Published
- 2020
- Full Text
- View/download PDF
10. Performance analysis of deep learning-based object detection algorithms on COCO benchmark: a comparative study.
- Author
-
Tian, Jiya, Jin, Qiangshan, Wang, Yizong, Yang, Jie, Zhang, Shuping, and Sun, Dengxun
- Subjects
OBJECT recognition (Computer vision) ,DEEP learning ,MACHINE learning ,ALGORITHMS ,SMART cities ,URBAN renewal - Abstract
This paper thoroughly explores the role of object detection in smart cities, specifically focusing on advancements in deep learning-based methods. Deep learning models gain popularity for their autonomous feature learning, surpassing traditional approaches. Despite progress, challenges remain, such as achieving high accuracy in urban scenes and meeting real-time requirements. The study aims to contribute by analyzing state-of-the-art deep learning algorithms, identifying accurate models for smart cities, and evaluating real-time performance using the Average Precision at Medium Intersection over Union (IoU) metric. The reported results showcase various algorithms' performance, with Dynamic Head (DyHead) emerging as the top scorer, excelling in accurately localizing and classifying objects. Its high precision and recall at medium IoU thresholds signify robustness. The paper suggests considering the mean Average Precision (mAP) metric for a comprehensive evaluation across IoU thresholds, if available. Despite this, DyHead stands out as the superior algorithm, particularly at medium IoU thresholds, making it suitable for precise object detection in smart city applications. The performance analysis using Average Precision at Medium IoU is reinforced by the Average Precision at Low IoU (APL), consistently depicting DyHead's superiority. These findings provide valuable insights for researchers and practitioners, guiding them toward employing DyHead for tasks prioritizing accurate object localization and classification in smart cities. Overall, the paper navigates through the complexities of object detection in urban environments, presenting DyHead as a leading solution with robust performance metrics. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. CNN-VAE: An intelligent text representation algorithm.
- Author
-
Xu, Saijuan, Guo, Canyang, Zhu, Yuhan, Liu, Genggeng, and Xiong, Neal
- Subjects
CONVOLUTIONAL neural networks ,BIG data ,MACHINE learning ,POLYSEMY ,SUPPORT vector machines ,K-nearest neighbor classification ,ALGORITHMS - Abstract
Collecting and analyzing data from all devices to improve the efficiency of business processes is an important task of Industrial Internet of Things (IIoT). In the age of data explosion, extensive text data generated by the IIoT have given birth to a variety of text representation methods. The task of text representation is to convert the natural language to a form that computer can understand with retaining the original semantics. However, these methods are difficult to effectively extract the semantic features among words and distinguish polysemy in natural language. Combining the advantages of convolutional neural network (CNN) and variational autoencoder (VAE), this paper proposes an intelligent CNN-VAE text representation algorithm as an advanced learning method for social big data within next-generation IIoT, which help users identify the information collected by sensors and perform further processing. This method employs the convolution layer to capture the local features of the context and uses the variational technique to reconstruct feature space to make it conform to the normal distribution. In addition, the improved word2vec model based on topical word embedding (TWE) is utilized to add topical information to word vectors to distinguish polysemy. This paper takes the social big data as an example to illustrate the way of the proposed algorithm applied in the next-generation IIoT and utilizes Cnews dataset to verify the performance of proposed method with four evaluating metrics (i.e., recall, accuracy, precision, and F1-score). Experimental results indicate that the proposed method outperforms word2vec-avg and CNN-AE in K-nearest neighbor (KNN), random forest (RF), and support vector machine (SVM) classifiers and distinguishes polysemy effectively. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
12. Discriminative shapelet learning via temporal clustering and matrix factorization.
- Author
-
Chen, Bo, Fang, Min, and Wang, GuiZhi
- Subjects
MACHINE learning ,MATRIX decomposition ,TIME series analysis ,CLASSIFICATION ,ALGORITHMS - Abstract
Identifying discriminative patterns, known as shapelets, within time series is a critical step in many time series classification tasks. A major limitation of shapelet learning is that often hindered by their unsupervised methods, treating shapelet learning as an unsupervised subsequence clustering process and discovery based on pre-defined metric, which performed sequentially. This sequential procedure presents challenges, as it fails to establish a direct connection between shapelets and samples, and lacks the capacity to explicitly incorporate label information. In this paper, we proposed a novel shapelet learning algorithm called Discriminative Shapelet Learning via Temporal Clustering and Matrix Factorization (DSLMF). DSLMF introduced a joint framework that combines matrix factorization and coherent temporal clustering to discovery salient and coherent feature subsets. To further enhance discriminability and prevent arbitrary shapelet shapes, DSLMF integrates a label-specific shapelet regularization as a guiding mechanism enabling the learning of shapelets optimized for higher classification performance. The proposed algorithm has shown to be effective for capturing the temporal cluster structure and interpretability of shapelet-based method. The results of experiments showcased in this paper highlight DSLMF's effectiveness in capturing temporal cluster structures and learning meaningful shapelets, ultimately leading to promising performance on benchmark datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Multimodal machine learning in precision health: A scoping review.
- Author
-
Kline, Adrienne, Wang, Hanyin, Li, Yikuan, Dennis, Saya, Hutch, Meghan, Xu, Zhenxing, Wang, Fei, Cheng, Feixiong, and Luo, Yuan
- Subjects
ONLINE information services ,NEUROLOGY ,SYSTEMATIC reviews ,MACHINE learning ,INDIVIDUALIZED medicine ,LITERATURE reviews ,MEDLINE ,HEALTH equity ,ALGORITHMS ,ONCOLOGY - Abstract
Machine learning is frequently being leveraged to tackle problems in the health sector including utilization for clinical decision-support. Its use has historically been focused on single modal data. Attempts to improve prediction and mimic the multimodal nature of clinical expert decision-making has been met in the biomedical field of machine learning by fusing disparate data. This review was conducted to summarize the current studies in this field and identify topics ripe for future research. We conducted this review in accordance with the PRISMA extension for Scoping Reviews to characterize multi-modal data fusion in health. Search strings were established and used in databases: PubMed, Google Scholar, and IEEEXplore from 2011 to 2021. A final set of 128 articles were included in the analysis. The most common health areas utilizing multi-modal methods were neurology and oncology. Early fusion was the most common data merging strategy. Notably, there was an improvement in predictive performance when using data fusion. Lacking from the papers were clear clinical deployment strategies, FDA-approval, and analysis of how using multimodal approaches from diverse sub-populations may improve biases and healthcare disparities. These findings provide a summary on multimodal data fusion as applied to health diagnosis/prognosis problems. Few papers compared the outputs of a multimodal approach with a unimodal prediction. However, those that did achieved an average increase of 6.4% in predictive accuracy. Multi-modal machine learning, while more robust in its estimations over unimodal methods, has drawbacks in its scalability and the time-consuming nature of information concatenation. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
14. Risk decision analysis of commercial insurance based on neural network algorithm.
- Author
-
Wang, Shanshan and Zhao, Zhenwang
- Subjects
BUSINESS insurance ,DECISION making ,RISK assessment ,ACTUARIAL risk ,ALGORITHMS - Abstract
To improve the effect of commercial insurance risk decision, this paper applies neural network algorithms to commercial insurance risk decision under the guidance of machine learning ideas, and selects the neural network algorithm based on the actual situation. Moreover, this paper analyzes the nature of risks of commercial insurance, analyzes the types of risks and risk relevance, constructs a commercial insurance risk decision model based on neural network algorithms, and determines the system process. In addition, this paper uses a combination method of qualitative and quantitative to identify the influencing factors of risk estimation to obtain relevant influencing factors, and verify the model proposed in this paper in combination with experimental research. From the experimental research results, it can be seen that the commercial insurance risk decision system based on neural network algorithm is very good in terms of decision effect. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
15. Trend analysis of global usage of digital soil mapping models in the prediction of potentially toxic elements in soil/sediments: a bibliometric review.
- Author
-
Agyeman, Prince Chapman, Ahado, Samuel Kudjo, Borůvka, Luboš, Biney, James Kobina Mensah, Sarkodie, Vincent Yaw Oppong, Kebonye, Ndiye M., and Kingsley, John
- Subjects
DIGITAL soil mapping ,TREND analysis ,PREDICTION models ,GLOBAL analysis (Mathematics) ,SOIL pollution - Abstract
The rising and continuous pollution of the soil from anthropogenic activities is of great concern. Owing to this concern, the advent of digital soil mapping (DSM) has been a tool that soil scientists use in this era to predict the potentially toxic element (PTE) content in the soil. The purpose of this paper was to conduct a review of articles, summarize and analyse the spatial prediction of potentially toxic elements, determine and compare the models' usage as well as their performance over time. Through Scopus, the Web of Science and Google Scholar, we collected papers between the year 2001 and the first quarter of 2019, which were tailored towards the spatial PTE prediction using DSM approaches. The results indicated that soil pollution emanates from diverse sources. However, it provided reasons why the authors investigate a piece of land or area, highlighting the uncertainties in mapping, number of publications per journal and continental efforts to research as well as published on trending issues regarding DSM. This paper reveals the complementary role machine learning algorithms and the geostatistical models play in DSM. Nevertheless, geostatistical approaches remain the most preferred model compared to machine learning algorithms. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
16. NICE: an algorithm for nearest instance counterfactual explanations.
- Author
-
Brughmans, Dieter, Leyman, Pieter, and Martens, David
- Subjects
MACHINE learning ,ALGORITHMS ,EXPLANATION ,CLASSIFICATION - Abstract
In this paper we propose a new algorithm, named NICE, to generate counterfactual explanations for tabular data that specifically takes into account algorithmic requirements that often emerge in real-life deployments: (1) the ability to provide an explanation for all predictions, (2) being able to handle any classification model (also non-differentiable ones), (3) being efficient in run time, and (4) providing multiple counterfactual explanations with different characteristics. More specifically, our approach exploits information from a nearest unlike neighbor to speed up the search process, by iteratively introducing feature values from this neighbor in the instance to be explained. We propose four versions of NICE, one without optimization and, three which optimize the explanations for one of the following properties: sparsity, proximity or plausibility. An extensive empirical comparison on 40 datasets shows that our algorithm outperforms the current state-of-the-art in terms of these criteria. Our analyses show a trade-off between on the one hand plausibility and on the other hand proximity or sparsity, with our different optimization methods offering users the choice to select the types of counterfactuals that they prefer. An open-source implementation of NICE can be found at https://github.com/ADMAntwerp/NICE. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. DeepEPhishNet: a deep learning framework for email phishing detection using word embedding algorithms.
- Author
-
Somesha, M and Pais, Alwyn Roshan
- Subjects
DEEP learning ,PHISHING ,SOCIAL engineering (Fraud) ,ARTIFICIAL neural networks ,EMAIL ,ALGORITHMS ,MACHINE learning - Abstract
Email phishing is a social engineering scheme that uses spoofed emails intended to trick the user into disclosing legitimate business and personal credentials. Many phishing email detection techniques exist based on machine learning, deep learning, and word embedding. In this paper, we propose a new technique for the detection of phishing emails using word embedding (Word2Vec, FastText, and TF-IDF) and deep learning techniques (DNN and BiLSTM network). Our proposed technique makes use of only four header based (From, Returnpath, Subject, Message-ID) features of the emails for the email classification. We applied several word embeddings for the evaluation of our models. From the experimental evaluation, we observed that the DNN model with FastText-SkipGram achieved an accuracy of 99.52% and BiLSTM model with FastText-SkipGram achieved an accuracy of 99.42%. Among these two techniques, DNN outperformed BiLSTM using the same word embedding (FastText-SkipGram) techniques with an accuracy of 99.52%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. A novel interpretable predictive model based on ensemble learning and differential evolution algorithm for surface roughness prediction in abrasive water jet polishing.
- Author
-
Xie, Shutong, He, Zongbao, Loh, Yee Man, Yang, Yu, Liu, Kunhong, Liu, Chao, Cheung, Chi Fai, Yu, Nan, and Wang, Chunjin
- Subjects
DIFFERENTIAL evolution ,WATER jets ,PREDICTION models ,MACHINE learning ,ABRASIVES ,ALGORITHMS ,GRINDING & polishing ,SURFACE roughness - Abstract
As an important indicator of the surface quality of workpieces, surface roughness has a great impact on production costs and the quality performance of the finished components. Effective surface roughness prediction can not only increase productivity but also reduce costs. However, the current methods for surface roughness prediction have some limitations. On the one hand, the prediction accuracy of classical experimental and statistical-based surface roughness prediction methods is low. On the other hand, the results of deep learning-based surface roughness prediction methods are uninterpretable due to their black-box learning mechanism. Therefore, this paper presents an ensemble learning with a differential evolution algorithm, applies it to the prediction of surface roughness of abrasive water jet polishing (AWJP), and conducts an interpretability analysis to identify key factors contributing to the prediction accuracy of surface roughness. First, we proposed automatically constructing features by an Evolution Forest algorithm to train the base regression models. The differential evolution algorithm with a simplified encoding mechanism was then used to search for the best weighted-ensemble to integrate the base regression models for obtaining highly accurate prediction results. Extensive experiments have been conducted on AWJP to validate the effectiveness of our proposed methods. The results show that the prediction accuracy of our proposed method is higher than the existing machine learning algorithms. In addition, this is the first of its time for the contributions of machining parameters (i.e., features) on surface roughness prediction by using interpretable analysis methods. The analysis results can provide a reference basis for subsequent experiments and studies. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Three-layer data center-based intelligent slice admission control algorithm for C-RAN using approximate reinforcement learning.
- Author
-
Khani, Mohsen, Jamali, Shahram, and Sohrabi, Mohammad Karim
- Subjects
MACHINE learning ,RADIO access networks ,5G networks ,ALGORITHMS ,REINFORCEMENT learning - Abstract
C-RAN (Cloud Radio Access Network) is a 5G architecture that consists of sites and three-layer Data Centers (DCs), which include the central office DC, local DC, and regional DC. Network slicing, which enables infrastructure providers (InP) to create independent logical networks, is essential in this architecture. By utilizing this technology, InPs can maximize the utility of the network by providing slices to service providers in response to their slice requests. However, almost all of the recent research on slice admission control (SAC) schemes has only considered one or two layers of DCs, which limits the efficiency of the slicing process and decreases network utility. To address these issues, this paper proposes an intelligent SAC scheme called ISAC that considers all three-layer DCs. Instead of relying on reinforcement learning algorithms like Q-learning, which are effective in discrete environments with limited state space but give poor performance in continuous environments, ISAC employs the Approximate Reinforcement Learning (ARL) algorithm. ARL is better suited for 5G network modeling because it can adapt to continuous environments, allowing for a more accurate representation of the underlying physical processes. Extensive simulation studies demonstrate that ISAC significantly improves performance in terms of slice request rejection rates, InP revenue, accepting more slices, and optimizing resource utilization. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
20. Maximizing intrusion detection efficiency for IoT networks using extreme learning machine.
- Author
-
Altamimi, Shahad and Abu Al-Haija, Qasem
- Subjects
MACHINE learning ,SUPERVISED learning ,INTERNET of things ,TELECOMMUNICATION systems ,CYBERTERRORISM ,ALGORITHMS - Abstract
Intrusion Detection Systems (IDSs) are crucial for safeguarding modern IoT communication networks against cyberattacks. IDSs must exhibit exceptional performance, low false positive rates, and significant flexibility in constructing attack patterns to efficiently identify and neutralize these attacks. This research paper discusses the use of an Extreme Learning Machine (ELM) as a new technique to enhance the performance of IDSs. The study utilizes two standard IDS-based IoT network datasets: NSL-KDD 2009 via Distilled-Kitsune 2021. Both datasets are used to assess the effectiveness of ELM in a conventional supervised learning setting. The study investigates the capacity of the ELM algorithm to handle high-dimensional and unbalanced data, indicating the potential to enhance IDS accuracy and efficiency. The research also examines the setup of ELM for both NSL_KDD and Kitsune using Python and Google COLAB to do binary and multi-class classification. The experimental evaluation revealed the proficient performance of the proposed ELM-based IDS among other implemented supervised learning-based IDSs and other state-of-the-art models in the same study area. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Stabilization of parareal algorithms for long-time computation of a class of highly oscillatory Hamiltonian flows using data.
- Author
-
Fang, Rui and Tsai, Richard
- Subjects
HAMILTON'S principle function ,HAMILTONIAN graph theory ,ALGORITHMS ,EIGENFUNCTIONS ,MULTISCALE modeling ,HAMILTONIAN systems ,PROBLEM solving ,PROOF of concept - Abstract
Applying parallel-in-time algorithms to multiscale Hamiltonian systems to obtain stable long-time simulations is very challenging. In this paper, we present novel data-driven methods aimed at improving the standard parareal algorithm developed by Lions et al. in 2001, for multiscale Hamiltonian systems. The first method involves constructing a correction operator to improve a given inaccurate coarse solver through solving a Procrustes problem using data collected online along parareal trajectories. The second method involves constructing an efficient, high-fidelity solver by a neural network trained with offline generated data. For the second method, we address the issues of effective data generation and proper loss function design based on the Hamiltonian function. We show proof-of-concept by applying the proposed methods to a Fermi-Pasta-Ulam (FPU) problem. The numerical results demonstrate that the Procrustes parareal method is able to produce solutions that are more stable in energy compared to the standard parareal. The neural network solver can achieve comparable or better runtime performance compared to numerical solvers of similar accuracy. When combined with the standard parareal algorithm, the improved neural network solutions are slightly more stable in energy than the improved numerical coarse solutions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. A metaheuristic-based task offloading scheme with a trade-off between delay and resource utilization in IoT platform.
- Author
-
Kumari, Nidhi and Jana, Prasanta K.
- Subjects
INTERNET of things ,MACHINE learning ,ALGORITHMS - Abstract
Fog computing has emerged as the most popular technology for processing delay-sensitive tasks in the Internet of Things platform. However, offloading tasks to suitable fog nodes (FNs) for their processing is a very challenging problem which is known to be NP-hard. In this paper, we propose a metaheuristic-based task offloading scheme that jointly optimizes the offloading delay and resource utilization. We first formulate the task offloading as a multi-objective optimization problem and then propose an algorithm based on the Discrete Jaya Algorithm (DJA) to solve it. The DJA is presented with efficient candidate representation and mutation operation. We also propose a task migration algorithm to transfer the partially completed task to another FN before failure. Through extensive simulations, we demonstrate the efficacy of the proposed work by comparing the state-of-the-art algorithms and validating it through the Friedman test. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. PAKDD'12 best paper: generating balanced classifier-independent training samples from unlabeled data.
- Author
-
Park, Youngja, Qi, Zijie, Chari, Suresh, and Molloy, Ian
- Subjects
PROBLEM solving ,DISTRIBUTION (Probability theory) ,INFORMATION theory ,KNOWLEDGE management ,ALGORITHMS ,ITERATIVE methods (Mathematics) - Abstract
We consider the problem of generating balanced training samples from an unlabeled data set with an unknown class distribution. While random sampling works well when the data are balanced, it is very ineffective for unbalanced data. Other approaches, such as active learning and cost-sensitive learning, are also suboptimal as they are classifier-dependent and require misclassification costs and labeled samples, respectively. We propose a new strategy for generating training samples, which is independent of the underlying class distribution of the data and the classifier that will be trained using the labeled data. Our methods are iterative and can be seen as variants of active learning, where we use semi-supervised clustering at each iteration to perform biased sampling from the clusters. We provide several strategies to estimate the underlying class distributions in the clusters and to increase the balancedness in the training samples. Experiments with both highly skewed and balanced data from the UCI repository and a private data set show that our algorithm produces much more balanced samples than random sampling or uncertainty sampling. Further, our sampling strategy is substantially more efficient than active learning methods. The experiments also validate that, with more balanced training data, classifiers trained with our samples outperform classifiers trained with random sampling or active learning. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
24. Systematic review of content analysis algorithms based on deep neural networks.
- Author
-
Rezaeenour, Jalal, Ahmadi, Mahnaz, Jelodar, Hamed, and Shahrooei, Roshan
- Subjects
ARTIFICIAL neural networks ,DEEP learning ,MACHINE learning ,INFORMATION technology ,NATURAL language processing ,ALGORITHMS - Abstract
Today according to social media, the internet, Etc. Data is rapidly produced and occupies a large space in systems that have resulted in enormous data warehouses; the progress in information technology has significantly increased the speed and ease of data flow.text mining is one of the most important methods for extracting a useful model through extracting and adapting knowledge from data sets. However, many studies have been conducted based on the usage of deep learning for text processing and text mining issues.The idea and method of text mining are one of the fields that seek to extract useful information from unstructured textual data that is used very today. Deep learning and machine learning techniques in classification and text mining and their type are discussed in this paper as well. Neural networks of various kinds, namely, ANN, RNN, CNN, and LSTM, are the subject of study to select the best technique. In this study, we conducted a Systematic Literature Review to extract and associate the algorithms and features that have been used in this area. Based on our search criteria, we retrieved 130 relevant studies from electronic databases between 1997 and 2021; we have selected 43 studies for further analysis using inclusion and exclusion criteria in Section 3.2. According to this study, hybrid LSTM is the most widely used deep learning algorithm in these studies, and SVM in machine learning method high accuracy in result shown. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
25. Cognitive spectrum sensing algorithm based on an RBF neural network and machine learning.
- Author
-
Yang, Shi and Tong, Chaoran
- Subjects
DEEP learning ,MACHINE learning ,ALGORITHMS ,IMAGE recognition (Computer vision) ,PARALLEL processing ,AUTODIDACTICISM - Abstract
After 70 years of intricate development, machine learning, represented by deep learning, is based on the multilevel structure of the human brain and the layer-by-layer analysis and processing mechanism of neuron connection and interaction information. The powerful parallel information processing ability of self-adaptation and self-learning has allowed for breakthroughs in many fields, among which the most representative is image recognition. Therefore, this paper proposed optimizing the RBF algorithm with machine learning (ML) to improve the recognition rate of spectrum sensing. The results showed that the average detection success rates of the RBF algorithm were 93.62%, 95.07%, 96.91%, 98.78% and 99.37% when the SNRs were − 8 dB, − 4 dB, 0 dB, 4 dB and 8 dB, respectively, and the other conditions were kept the same. The average detection success rates of the SVM/RBF algorithm were 97.65%, 99.63%, 99.76%, 99.91% and 99.88%, respectively. The average detection success rate of the SVM/RBF algorithm was significantly higher than that of the RBF algorithm. This indicates that analyzing the RBF neural network algorithm through ML can improve the success rate of spectrum sensing, which highlights a new direction for the application of ML and neural networks. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. Non-smooth setting of stochastic decentralized convex optimization problem over time-varying Graphs.
- Author
-
Lobanov, Aleksandr, Veprikov, Andrew, Konin, Georgiy, Beznosikov, Aleksandr, Gasnikov, Alexander, and Kovalev, Dmitry
- Subjects
MACHINE learning ,ALGORITHMS ,NONSMOOTH optimization - Abstract
Distributed optimization has a rich history. It has demonstrated its effectiveness in many machine learning applications, etc. In this paper we study a subclass of distributed optimization, namely decentralized optimization in a non-smooth setting. Decentralized means that m agents (machines) working in parallel on one problem communicate only with the neighbors agents (machines), i.e. there is no (central) server through which agents communicate. And by non-smooth setting we mean that each agent has a convex stochastic non-smooth function, that is, agents can hold and communicate information only about the value of the objective function, which corresponds to a gradient-free oracle. In this paper, to minimize the global objective function, which consists of the sum of the functions of each agent, we create a gradient-free algorithm by applying a smoothing scheme via l 2 randomization. We also verify in experiments the obtained theoretical convergence results of the gradient-free algorithm proposed in this paper. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
27. Deep learning for the security of software-defined networks: a review.
- Author
-
Taheri, Roya, Ahmed, Habib, and Arslan, Engin
- Subjects
DEEP learning ,SOFTWARE-defined networking ,COMPUTER network security ,CYBERTERRORISM ,MACHINE learning ,ALGORITHMS - Abstract
As the scale and complexity of networks grow rapidly, management, maintenance, and optimization of them are becoming increasingly challenging tasks for network administrators. Software-Defined Networking (SDN) was introduced to facilitate these tasks as it offers logically centralized control, a global view of the network, and software-based traffic analysis, thus, it is widely adopted of SDN to manage large-scale networks. On the other hand, SDN is not immune to cyber attacks. In fact, its centralized architecture makes it more vulnerable to certain types of attacks, such as denial of service. Various attack mitigation strategies are proposed to strengthen the security of SDNs including statistical, threshold-based, and Machine Learning (ML) methods. Among them, Deep Learning (DL)-based models attained the best results as they were able to extract the complex relationship between input parameters and output that could not be achieved with other solutions. Hence, this paper presents a comprehensive survey of the literature on the utilization of different DL algorithms for the security of SDN. We first explain the types of attacks that SDNs are exposed to, then present papers that applied DL to detect and/or mitigate these attacks. We further discuss the public datasets used to train DL models and evaluate their advantages and disadvantages. Finally, we share insights into future research directions to improve the efficiency of DL methods for SDN security. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
28. Performance analysis on dictionary learning and sparse representation algorithms.
- Author
-
Ng, Suit Mun, Yazid, Haniza, and Mustafa, Nazahah
- Subjects
SPARSE approximations ,ALGORITHMS ,IMAGE representation ,MACHINE learning ,SIGNAL-to-noise ratio ,REPRESENTATION theory - Abstract
Theoretically, the Super-Resolution (SR) reconstruction scheme is a method which is performed by many applications nowadays for the purpose of generating a High-Resolution (HR) image using the input Low-Resolution (LR) images by filling in the missing high frequency information. In addition, the SR reconstruction implemented based on the theory of sparse representation techniques is known as an effective way to produce HR images using images patches generated from the LR images. In order to improve the quality of denoised images produced by using the sparse representation techniques, a scheme called dictionary learning algorithms could be considered. Thus, the objective of this paper is to provide a performance comparison on the effectiveness of applying the dictionary learning steps with sparse representation algorithms in producing a better denoised image. In this case, the average Peak Signal-to-Noise ratio (PSNR) and Structural Similarity Index Metric (SSIM) values of the denoised image obtained by using Algorithms 1, 2, and 3 which combined the use of dictionary learning and sparse representation algorithms were compared with the values obtained from images produced by applying only sparse regularisation methods. As a conclusion, the denoised images produced by Algorithm 1 in this paper had the greatest average PSNR and SSIM values. Hence, the algorithm with the implementation of the dictionary learning process with sparse representation methods is able to achieve a better result in enhancing the low-resolution images. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
29. A distributed adaptive policy gradient method based on momentum for multi-agent reinforcement learning.
- Author
-
Shi, Junru, Wang, Xin, Zhang, Mingchuan, Liu, Muhua, Zhu, Junlong, and Wu, Qingtao
- Subjects
REINFORCEMENT learning ,MACHINE learning ,SAMPLING (Process) ,CHANGE agents ,ALGORITHMS - Abstract
Policy Gradient (PG) method is one of the most popular algorithms in Reinforcement Learning (RL). However, distributed adaptive variants of PG are rarely studied in multi-agent. For this reason, this paper proposes a distributed adaptive policy gradient algorithm (IS-DAPGM) incorporated with Adam-type updates and importance sampling technique. Furthermore, we also establish the theoretical convergence rate of O (1 / T) , where T represents the number of iterations, it can match the convergence rate of the state-of-the-art centralized policy gradient methods. In addition, many experiments are conducted in a multi-agent environment, which is a modification on the basis of Particle world environment. By comparing with some other distributed PG methods and changing the number of agents, we verify the performance of IS-DAPGM is more efficient than the existing methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Label distribution feature selection based on label-specific features.
- Author
-
Shu, Wenhao, Xia, Qiang, and Qian, Wenbin
- Subjects
MACHINE learning ,COLLABORATIVE learning ,STATISTICAL correlation ,ALGORITHMS ,AMBIGUITY ,FEATURE selection - Abstract
Label distribution learning, where deal with label ambiguity by describing the degree of relevance of each label to a specific instance. As a novel machine learning paradigm, the curse of dimensionality is one of the prominent problems. Feature selection is a vital preprocessing step to reduce the high dimensionality of data. However, most existing label distribution feature selection methods focus on selecting a feature subset that has relevant capabilities for all labels, ignoring label-specific features with the maximum discriminatory power for each label. To tackle this issue, a label distribution feature selection algorithm based on label-specific features is proposed in this paper. Initially, we introduce a feature selection optimization model for label distribution data that simultaneously considers common and label-specific features, leveraging sparse learning to further investigate the intricate relationships between features and labels. Subsequently, the label correlation coefficient is employed to enhance the collaborative learning effect of labels. Finally, the relevance between features and labels is taken into account to guide the feature selection process, which can effectively eliminate the redundant features. Comprehensive experiments demonstrate the advantage of our proposed method over other well-established feature selection algorithms for selecting label-specific features to label distribution data. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. Non-iterative approaches in training feed-forward neural networks and their applications.
- Author
-
Wang, Xizhao and Cao, Weipeng
- Subjects
FEEDFORWARD neural networks ,DEEP learning ,ALGORITHMS ,PROBLEM solving ,MACHINE learning - Abstract
Focusing on non-iterative approaches in training feed-forward neural networks, this special issue includes 12 papers to share the latest progress, current challenges, and potential applications of this topic. This editorial presents a background of the special issue and a brief introduction to the 12 contributions. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
32. A systematic literature review on hardware implementation of artificial intelligence algorithms.
- Author
-
Talib, Manar Abu, Majzoub, Sohaib, Nasir, Qassim, and Jamal, Dina
- Subjects
ARTIFICIAL intelligence ,MACHINE learning ,ALGORITHMS ,HARDWARE ,GRAPHICS processing units - Abstract
Artificial intelligence (AI) and machine learning (ML) tools play a significant role in the recent evolution of smart systems. AI solutions are pushing towards a significant shift in many fields such as healthcare, autonomous airplanes and vehicles, security, marketing customer profiling and other diverse areas. One of the main challenges hindering the AI potential is the demand for high-performance computation resources. Recently, hardware accelerators are developed in order to provide the needed computational power for the AI and ML tools. In the literature, hardware accelerators are built using FPGAs, GPUs and ASICs to accelerate computationally intensive tasks. These accelerators provide high-performance hardware while preserving the required accuracy. In this work, we present a systematic literature review that focuses on exploring the available hardware accelerators for the AI and ML tools. More than 169 different research papers published between the years 2009 and 2019 are studied and analysed. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
33. Seeing like an algorithm: the limits of using remote sensing to link vessel movements with worker abuse at sea.
- Author
-
Rudolph, Terence Adam
- Subjects
ABUSE of employees ,REMOTE sensing ,MACHINE learning ,GEOGRAPHIC information systems ,GEOSPATIAL data ,MARITIME boundaries ,SHIP captains ,ALGORITHMS - Abstract
The ship tracking and mapping capabilities that geospatial technology provides create an opportunity to observe fishing vessels as they move through established maritime boundaries. This paper connects data availability to ground-truthing research and explores the limits of vessel movement mapping in representing worker abuse at sea through three related themes. First, a conceptual background links the advancements in maritime remote sensing to critical GIS scholarship and provides a background on worker abuse aboard Taiwanese fishing vessels. Second, the paper examines the potential of machine learning algorithms to represent worker abuse at sea, arguing that more extensive ground-truthing research with workers could help address variations in the data and limited data sets. Third, I use remote sensing data to identify and unpack Taiwanese fishing across the three EEZs with the most concentrated Taiwanese fishing activity: starting with Taiwan, followed by the Falkland Islands, and Seychelles. I argue that fishing activity and the digital representation of vessel movements are governed by terrestrial geopolitics and subject to manipulation by ship captains. Finally, the conclusion offers recommendations for how future research can capitalize on the capabilities of AIS, particularly with respect to addressing problems of working conditions and abuse at sea. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
34. Weak aggregating specialist algorithm for online portfolio selection.
- Author
-
He, Jin'an, Yin, Shicheng, and Peng, Fangping
- Subjects
ONLINE algorithms ,MACHINE learning ,ONLINE education ,FINANCIAL markets ,ALGORITHMS - Abstract
This paper proposes a novel online learning algorithm, named weak aggregating specialist algorithm (WASA), and presents its theoretical bound. This algorithm has a flexible feature, which is to allow abandoning some expert advice according to pre-set rules. Based on this algorithm, a new online portfolio strategy named weak aggregating specialized CRP (WASC) is designed, which only aggregates awake specialized expert advice. Firstly, a pool of special constant rebalanced portfolios CRP (b) strategies is employed to construct the index set of specialized experts. Secondly, a distance function is exploited to measure the distance between the current adjusted portfolio and each specialized expert advice, and the index set of awake specialized experts is constructed. Finally, the portfolio is updated by aggregating all awake specialized expert advice. Furthermore, theoretical and experimental analyses are established to illustrate the performance of the proposed strategy WASC. Theoretical results guarantee that WASC performs as well as the best specialized expert. Experimental results show that WASC outperforms some existing strategies in terms of the return and risk metrics, which illustrates its effectiveness in various real financial markets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Performance Comparison of Different HTM-Spatial Pooler Algorithms Based on Information-Theoretic Measures.
- Author
-
Sanati, Shiva, Rouhani, Modjtaba, and Hodtani, Ghosheh Abed
- Subjects
MACHINE learning ,SHORT-term memory ,LONG-term memory ,SEQUENTIAL learning ,ALGORITHMS ,DISTRIBUTED algorithms ,BOOSTING algorithms - Abstract
Hierarchical temporal memory (HTM) is a promising unsupervised machine-learning algorithm that models key principles of neocortical computation. One of the main components of HTM is the spatial pooler (SP), which encodes binary input streams into sparse distributed representations (SDRs). In this paper, we propose an information-theoretic framework for the performance comparison of HTM-spatial pooler (SP) algorithms, specifically, for quantifying the similarities and differences between sparse distributed representations in SP algorithms. We evaluate SP's standalone performance, as well as HTM's overall performance. Our comparison of various SP algorithms using Renyi mutual information, Renyi divergence, and Henze–Penrose divergence measures reveals that the SP algorithm with learning and a logarithmic boosting function yields the most effective and useful data representation. Moreover, the most effective SP algorithm leads to superior HTM results. In addition, we utilize our proposed framework to compare HTM with other state-of-the-art sequential learning algorithms. We illustrate that HTM exhibits superior adaptability to pattern changes over time than long short term memory (LSTM), gated recurrent unit (GRU) and online sequential extreme learning machine (OS-ELM) algorithms. This superiority is evident from the lower Renyi divergence of HTM (0.23) compared to LSTM6000 (0.33), LSTM3000 (0.38), GRU (0.41), and OS-ELM (0.49). HTM also achieved the highest Renyi mutual information value of 0.79, outperforming LSTM6000 (0.73), LSTM3000 (0.71), GRU (0.68), and OS-ELM (0.62). These findings not only confirm the numerous advantages of HTM over other sequential learning algorithm, but also demonstrate the effectiveness of our proposed information-theoretic approach as a powerful framework for comparing and evaluating various learning algorithms. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. Fedisp: an incremental subgradient-proximal-based ring-type architecture for decentralized federated learning.
- Author
-
Huang, Jianjun, Rui, Zihao, and Kang, Li
- Subjects
FEDERATED learning ,DATA privacy ,MACHINE learning ,DISTRIBUTED algorithms ,ITERATIVE learning control ,ALGORITHMS - Abstract
Federated learning (FL) represents a promising distributed machine learning paradigm for resolving data isolation due to data privacy concerns. Nevertheless, most vanilla FL algorithms, which depend on a server, encounter the problem of reliability and a high communication burden in real cases. Decentralized federated learning (DFL) that does not follow the star topology faces the challenges of weight divergence and inferior communication efficiency. In this paper, a novel DFL framework called federated incremental subgradient-proximal (FedISP) is proposed that utilizes the incremental method to perform model updates to alleviate weight divergence. In our setup, multiple clients are distributed in a ring topology and communicate in a cyclic manner, which significantly mitigates the communication load. A convergence guarantee is given under the convex condition to demonstrate the impact of the learning rate on our algorithms, which further improves the performance of FedISP. Extensive experiments on benchmark datasets validate the effectiveness of the proposed approach in both independent and identically distributed (IID) and non-IID settings while illustrating the advantages of the FedISP algorithm in achieving model consensus and saving communication costs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. VAMPIRE: vectorized automated ML pre-processing and post-processing framework for edge applications.
- Author
-
Daher, Ali W., Ferrari, Enrico, Muselli, Marco, Chible, Hussein, and Caviglia, Daniele D.
- Subjects
FEATURE extraction ,MACHINE learning ,VAMPIRES ,RASPBERRY Pi ,ARTIFICIAL intelligence ,REMOTE sensing - Abstract
Machine learning techniques aim to mimic the human ability to automatically learn how to perform tasks through training examples. They have proven capable of tasks such as prediction, learning and adaptation based on experience and can be used in virtually any scientific application, ranging from biomedical, robotic, to business decision applications, and others. However, the lack of domain knowledge for a particular application can make feature extraction ineffective or even unattainable. Furthermore, even in the presence of pre-processed datasets, the iterative process of optimizing Machine Learning parameters, which do not translate from one domain to another, maybe difficult for inexperienced practitioners. To address these issues, we present in this paper a Vectorized Automated ML Pre-processIng and post-pRocEssing framework, approximately named (VAMPIRE), which implements feature extraction algorithms capable of converting large time-series recordings into datasets. Also, it introduces a new concept, the Activation Engine, which is attached to the output of a Multi Layer Perceptron and extracts the optimal threshold to apply binary classification. Moreover, a tree-based algorithm is used to achieve multi-class classification using the Activation Engine. Furthermore, the internet of things gives rise to new applications such as remote sensing and communications, so consequently applying Machine Learning to improve operation accuracy, latency, and reliability is beneficial in such systems. Therefore, all classifications in this paper were performed on the edge in order to reach high accuracy with limited resources. Moreover, forecasts were applied on three unrelated biomedical datasets, and on two other pre-processed urban and activity detection datasets. Features were extracted when required, and training and testing were performed on the Raspberry Pi remotely, where high accuracy and inference speed were achieved in every experiment. Additionally, the board remained competitive in terms of power consumption when compared with a laptop which was optimized using a Graphical Processing Unit. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
38. Fast all zero block detection algorithm for versatile video coding.
- Author
-
Niu, Weihong, Huang, Xiaofeng, Yin, Haibing, Lu, Yu, Zhou, Yang, and Yan, Chenggang
- Subjects
VIDEO coding ,ALGORITHMS ,MACHINE learning - Abstract
The new generation versatile video coding (VVC) standard brings extremely high compression efficiency. Meanwhile, the complexity of the encoder is also greatly increased. In view of this, this paper proposes a fast algorithm for all zero block detection to speed up the quantization process, thereby reducing the complexity of the encoder. The method proposed in this paper consists of three parts. Firstly, genuine all zero blocks are detected based on a fixed threshold which is derived by hard decision quantization formula. Secondly, parts of pseudo all zero blocks are detected based on an adaptive threshold which is derived by analyzing the positions of transform coefficients under a certain condition. Finally, for the remaining pseudo all zero blocks, machine learning is introduced and the decision is made through the fully connected neural network. Experimental results show that the proposed fast algorithm achieves up to 7.505% and 7.049% coding time saving under Low Delay B and Random Access configurations with only 0.470% and 0.578% performance loss on average, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. DC-SHAP Method for Consistent Explainability in Privacy-Preserving Distributed Machine Learning.
- Author
-
Bogdanova, Anna, Imakura, Akira, and Sakurai, Tetsuya
- Subjects
MACHINE learning ,DEEP learning ,COMMERCIAL products ,ALGORITHMS ,ARTIFICIAL intelligence - Abstract
Ensuring the transparency of machine learning models is vital for their ethical application in various industries. There has been a concurrent trend of distributed machine learning designed to limit access to training data for privacy concerns. Such models, trained over horizontally or vertically partitioned data, present a challenge for explainable AI because the explaining party may have a biased view of background data or a partial view of the feature space. As a result, explanations obtained from different participants of distributed machine learning might not be consistent with one another, undermining trust in the product. This paper presents an Explainable Data Collaboration Framework based on a model-agnostic additive feature attribution algorithm (KernelSHAP) and Data Collaboration method of privacy-preserving distributed machine learning. In particular, we present three algorithms for different scenarios of explainability in Data Collaboration and verify their consistency with experiments on open-access datasets. Our results demonstrated a significant (by at least a factor of 1.75) decrease in feature attribution discrepancies among the users of distributed machine learning. The proposed method improves consistency among explanations obtained from different participants, which can enhance trust in the product and enable ethical application in various industries. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. A Structure-Aware Convolutional Neural Network for Automatic Diagnosis of Fungal Keratitis with In Vivo Confocal Microscopy Images.
- Author
-
Liang, Shanshan, Zhong, Jing, Zeng, Hongwei, Zhong, Peixun, Li, Saiqun, Liu, Huijun, and Yuan, Jin
- Subjects
IN vivo studies ,MICROSCOPY ,MACHINE learning ,FUNGAL keratitis ,RESEARCH funding ,ARTIFICIAL neural networks ,COMPUTER-aided diagnosis ,SENSITIVITY & specificity (Statistics) ,ALGORITHMS ,EYE diseases - Abstract
Fungal keratitis (FK) is a common and severe corneal disease, which is widely spread in tropical and subtropical areas. Early diagnosis and treatment are vital for patients, with confocal microscopy cornea imaging being one of the most effective methods for the diagnosis of FK. However, most cases are currently diagnosed by the subjective judgment of ophthalmologists, which is time-consuming and heavily depends on the experience of the ophthalmologists. In this paper, we introduce a novel structure-aware automatic diagnosis algorithm based on deep convolutional neural networks for the accurate diagnosis of FK. Specifically, a two-stream convolutional network is deployed, combining GoogLeNet and VGGNet, which are two commonly used networks in computer vision architectures. The main stream is used for feature extraction of the input image, while the auxiliary stream is used for feature discrimination and enhancement of the hyphae structure. Then, the features are combined by concatenating the channel dimension to obtain the final output, i.e., normal or abnormal. The results showed that the proposed method achieved accuracy, sensitivity, and specificity of 97.73%, 97.02%, and 98.54%, respectively. These results suggest that the proposed neural network could be a promising computer-aided FK diagnosis solution. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. Influencing factors of stock returns based on Fama–French model and intelligent algorithm.
- Author
-
Rui, Xunyuan
- Subjects
MACHINE learning ,STOCKS (Finance) ,INVESTORS ,INVESTMENT advisors ,ALGORITHMS - Abstract
Due to the large number of stock companies, complex stock categories and inconsistent evaluation standards of market value, it is not conducive to the choice of investors. Based on the theory and practice of traditional stock performance evaluation model, this paper integrates the algorithm thought structure of Fama–French five-factor model and proposes a machine learning algorithm model for stock performance research. In addition, this paper also builds a model that can evaluate the style and timing ability of fund managers to improve the fund performance evaluation system to a greater extent. With the help of the performance evaluation updating function of the model, it provides a new experience material in the empirical research composition of the fund performance field. In the system test module, the former prediction data and the actual experimental data are integrated, and the two are sorted out and compared, which proves the feasibility and effectiveness of the proposed algorithm model. The final experimental results verify the usefulness of the model in stock return prediction and analysis of influencing factors. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
42. Multi-attribute overlapping radar working pattern recognition based on K-NN and SVM-BP.
- Author
-
Liao, Yanping and Chen, Xinyu
- Subjects
PATTERN recognition systems ,RADAR ,ALGORITHMS ,TRACKING radar ,TRAINING planes - Abstract
A recognition model named the SVM-NP is proposed in this paper to address the multi-attribute overlap in radar working recognition. The model is based on the K-NN boundary preselection algorithm and SVM-BP algorithm. Traditional classifiers tend to neglect the overlap of samples' attributes in classification, which leads to the low accuracy of classifiers. The K-NN boundary preselection can quickly select boundary samples from the total ones and reduce the whole samples' attribute overlap. The SVM-BP algorithm is improved based on the SVM-RFE algorithm, and the boundary samples with high attribute overlap are divided into many planes for training and testing. Compared with traditional methods, the overlap of sample attributes can be reduced twice in this model. Theoretical analysis and experimental results verify that the model proposed in this paper displays better performance in classification when appropriate parameters are provided. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
43. Feature-label dual-mapping for missing label-specific features learning.
- Author
-
Zhang, Lulu, Cheng, Yusheng, Wang, Yibin, and Pei, Gensheng
- Subjects
MACHINE learning ,ALGORITHMS - Abstract
Label-specific features learning can effectively exploit the unique features of each label, which alleviates the high dimensionality and improves the classification performance of multi-label. However, most existing label-specific features learning algorithms assume that label space is complete, ignoring the effect of missing labels on the classification accuracy. Some methods try to recover the missing labels first and then learn the mapping between the completed label matrix and the feature matrix. However, early intervention in the recovery of missing labels may affect the distribution of original labels to a certain extent. In this paper, feature-label dual-mapping for missing label-specific features learning is proposed. According to the information that the label depends on the feature, the dual-mapping weight of the complete feature space and the missing label space is jointly learned. Therefore, the proposed algorithm is to conduct latent missing labels recovery by feature-label dual-mapping to directly obtain target weight in this paper, avoiding the negative influence of early label recovery intervention. Compared with several state-of-the-art methods in 10 benchmark multi-label data sets, the results show that the proposed algorithm is reasonable and effective. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
44. Research on path planning algorithm of mobile robot based on reinforcement learning.
- Author
-
Pan, Guoqian, Xiang, Yong, Wang, Xiaorui, Yu, Zhongquan, and Zhou, Xinzhi
- Subjects
MOBILE robots ,REINFORCEMENT learning ,MOBILE learning ,ALGORITHMS ,MACHINE learning ,PROBLEM solving - Abstract
In order to solve the problems of low learning efficiency and slow convergence speed when mobile robot uses reinforcement learning method for path planning in complex environment, a reinforcement learning method based on each round path planning result is proposed. Firstly, the algorithm adds obstacle learning matrix to improve the success rate of path planning; and introduces heuristic reward to speed up the learning process by reducing the search space; then proposes a method of dynamically adjusting the exploration factor to balance the exploration and utilization in path planning, so as to further improve the performance of the algorithm. Finally, the simulation experiment in grid environment shows that compared with Q-learning algorithm, the improved algorithm not only shortens the average path length of the robot to reach the target position, but also speeds up the learning efficiency of the algorithm, so that the robot can find the optimal path more quickly. The code of EPRQL algorithm proposed in this paper has been published to GitHub: https://github.com/panpanpanguoguoqian/mypaper1.git. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
45. A group interest-based collaborative filtering algorithm for multimedia information.
- Author
-
Wang, Tao
- Subjects
FACTORIZATION ,RECOMMENDER systems ,SOCIAL networks ,ALGORITHMS ,NONNEGATIVE matrices ,MACHINE learning ,COMPUTER software - Abstract
This paper proposes a collaborative filtering algorithm based on user group interest. A novel co-clustering method (BalClust) and various weighted non-negative matrix factorization algorithms are used in the proposed method. The BalClust method is used to divide the raw rating matrix into clusters, which are smaller than the original matrix. Then, the balance factor is introduced to consider the user weight and the item-based CF (collaborative filtering). To predict the rating of the unknown items in the cluster, the non-negative matrix factorization algorithm was used. The proposed method achieves higher predicting accuracy and efficiency on low dimensional and homogeneous sub-matrices, and the method also reduces the computational complexity by combining the user and item-based CF. Based on the proposed method, this paper proposed an incremental learning method to ensure data accuracy and timeliness to overcome the problem brought by data updates. The experimental results show the proposed methods outperformed traditional CF algorithms, and the completion time is reduced. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
46. Computational thematics: comparing algorithms for clustering the genres of literary fiction.
- Author
-
Sobchuk, Oleg and Šeļa, Artjoms
- Subjects
LITERARY form ,MACHINE learning ,ALGORITHMS ,THEMATIC analysis ,FEATURE extraction - Abstract
What are the best methods of capturing thematic similarity between literary texts? Knowing the answer to this question would be useful for automatic clustering of book genres, or any other thematic grouping. This paper compares a variety of algorithms for unsupervised learning of thematic similarities between texts, which we call "computational thematics". These algorithms belong to three steps of analysis: text pre-processing, extraction of text features, and measuring distances between the lists of features. Each of these steps includes a variety of options. We test all the possible combinations of these options. Every combination of algorithms is given a task to cluster a corpus of books belonging to four pre-tagged genres of fiction. This clustering is then validated against the "ground truth" genre labels. Such comparison of algorithms allows us to learn the best and the worst combinations for computational thematic analysis. To illustrate the difference between the best and the worst methods, we then cluster 5000 random novels from the HathiTrust corpus of fiction. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
47. Optimal selection of benchmarking datasets for unbiased machine learning algorithm evaluation.
- Author
-
Pereira, João Luiz Junho, Smith-Miles, Kate, Muñoz, Mario Andrés, and Lorena, Ana Carolina
- Subjects
MACHINE learning ,SUPERVISED learning ,METAHEURISTIC algorithms ,CLASSIFICATION algorithms ,ALGORITHMS - Abstract
Whenever a new supervised machine learning (ML) algorithm or solution is developed, it is imperative to evaluate the predictive performance it attains for diverse datasets. This is done in order to stress test the strengths and weaknesses of the novel algorithms and provide evidence for situations in which they are most useful. A common practice is to gather some datasets from public benchmark repositories for such an evaluation. But little or no specific criteria are used in the selection of these datasets, which is often ad-hoc. In this paper, the importance of gathering a diverse benchmark of datasets in order to properly evaluate ML models and really understand their capabilities is investigated. Leveraging from meta-learning studies evaluating the diversity of public repositories of datasets, this paper introduces an optimization method to choose varied classification and regression datasets from a pool of candidate datasets. The method is based on maximum coverage, circular packing, and the meta-heuristic Lichtenberg Algorithm for ensuring that diverse datasets able to challenge the ML algorithms more broadly are chosen. The selections were compared experimentally with a random selection of datasets and with clustering by k-medoids and proved to be more effective regarding the diversity of the chosen benchmarks and the ability to challenge the ML algorithms at different levels. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
48. Efficient improvement of energy detection technique in cognitive radio networks using K-nearest neighbour (KNN) algorithm.
- Author
-
Musuvathi, Aneesh Sarjit S., Archbald, Jofin F., Velmurugan, T., Sumathi, D., Renuga Devi, S., and Preetha, K. S.
- Subjects
COGNITIVE radio ,RADIO networks ,MACHINE learning ,WIRELESS channels ,ALGORITHMS ,RESOURCE allocation - Abstract
With the birth of the IoT era, it is evident that the existing number of devices is going to rise exponentially. Any two devices will communicate with each other using the same frequency band with limited availability. Therefore, it is of vital importance that this frequency band used for communication be used efficiently to accommodate the maximum number of devices with the available radio resources. Cognitive radio (CR) technology serves this exact purpose. The stated one is an intelligent radio that is made to automatically identify the optimal wireless channel in the available wireless spectrum at a given instant. An important functionality of CR is spectrum sensing. Energy detection is a very popular algorithm used for spectrum sensing in CR technology for efficient allocation of radio resources to the devices intended to communicate with each other. Energy detection detects the presence of a primary user (PU) signal by continuously monitoring a selected frequency bandwidth. The conventional energy detection technique is known to perform poorly in lower SNR ranges. This paper works towards the improvement of the energy detection algorithm with the help of machine learning (ML). The ML model uses the general properties of the signal as training data and classifies between a PU signal and noise at very low SNR ranges (− 25 to − 10 dB). In this research, a K-nearest neighbours (KNN) model is selected for its versatility and simplicity. Upon testing the model with an out-of-sample dataset, the KNN model produced a detection accuracy of 94.5%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. MBSSA-Bi-AESN: Classification prediction of bi-directional adaptive echo state network based on modified binary salp swarm algorithm and feature selection.
- Author
-
Wu, Xunjin, Zhan, Jianming, Li, Tianrui, Ding, Weiping, and Pedrycz, Witold
- Subjects
FEATURE selection ,SUBSET selection ,DEMAND forecasting ,ALGORITHMS ,MACHINE learning ,TIME series analysis ,CLASSIFICATION - Abstract
In the era of big data, the demand for multivariate time series prediction has surged, drawing increased attention to feature selection and neural networks in machine learning. However, certain feature selection methods neglect the alignment between actual data sample differences and clustering results, while neural networks lack automatic parameter adjustment in response to changing target features. This paper presents the MBSSA-Bi-AESN model, a Bi-directional Adaptive Echo State Network that utilizes the modified salp swarm algorithm (MBSSA) and feature selection to address the limitations of manually set parameters. Initial feature subset selection involves assigning weights based on the consistency of clustering results with differences. Subsequently, the four critical parameters in the Bi-AESN model are optimized using MBSSA. The optimized Bi-AESN model and selected feature subset are then integrated for simultaneous model learning and optimal feature subset selection. Experimental analysis on eight datasets demonstrates the superior prediction accuracy of the MBSSA-Bi-AESN model compared to benchmark models, underscoring its feasibility, validity, and universality. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. Review of State-of-the-Art Microwave Filter Tuning Techniques and Implementation of a Novel Tuning Algorithm Using Expert-Based Hybrid Learning.
- Author
-
Sekhri, Even, Kapoor, Rajiv, and Tamre, Mart
- Subjects
MICROWAVE filters ,BLENDED learning ,MACHINE learning ,ALGORITHMS ,DEEP learning - Abstract
Present-day demand and supply of connectivity necessitate the rapid production of Microwave (MW) filter units. The production of these filters is then followed by the step of utmost importance in the assembly line, viz., the 'tuning of the filter', as tuning is crucial to meeting the selectivity requirements of the band. Since the advent of filters, tuning has always been done manually, and hence it is considered a bottleneck by experts in the field. Thus, the need to automate the system is highly implied. The goal of the current work is to outline various MW filter tuning techniques that have been advocated by the community of researchers. The limitations of the said research works and their comparative analysis are also encapsulated in tabular form in the present paper. The paper ends with the implementation of an Expert-Based Hybrid Deep Learning Algorithm to fully automate the filter tuning process. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.