7,573 results
Search Results
2. Factors associating with or predicting more cited or higher quality journal articles: An Annual Review of Information Science and Technology (ARIST) paper.
- Author
-
Kousha, Kayvan and Thelwall, Mike
- Subjects
- *
ABSTRACTING , *PUBLISHING , *READABILITY (Literary style) , *SERIAL publications , *METADATA , *BIBLIOGRAPHY , *CONFERENCES & conventions , *REGRESSION analysis , *MACHINE learning , *CITATION analysis , *INFORMATION science , *BIBLIOGRAPHICAL citations , *INTERPROFESSIONAL relations , *PERIODICAL articles , *IMPACT factor (Citation analysis) , *INFORMATION technology , *ABSTRACTING & indexing services , *MEDICAL research - Abstract
Identifying factors that associate with more cited or higher quality research may be useful to improve science or to support research evaluation. This article reviews evidence for the existence of such factors in article text and metadata. It also reviews studies attempting to estimate article quality or predict long‐term citation counts using statistical regression or machine learning for journal articles or conference papers. Although the primary focus is on document‐level evidence, the related task of estimating the average quality scores of entire departments from bibliometric information is also considered. The review lists a huge range of factors that associate with higher quality or more cited research in some contexts (fields, years, journals) but the strength and direction of association often depends on the set of papers examined, with little systematic pattern and rarely any cause‐and‐effect evidence. The strongest patterns found include the near universal usefulness of journal citation rates, author numbers, reference properties, and international collaboration in predicting (or associating with) higher citation counts, and the greater usefulness of citation‐related information for predicting article quality in the medical, health and physical sciences than in engineering, social sciences, arts, and humanities. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Development of new correlation for the prediction of power number for closed clearance impellers using machine learning methods trained on literature data.
- Author
-
Joshi, Sumit S., Dalvi, Vishwanath H., Vitankar, Vivek S., Joshi, Jyeshtharaj B., and Joshi, Aniruddha J.
- Subjects
MACHINE learning ,ARTIFICIAL neural networks ,NEWTONIAN fluids ,PAPER pulp ,REYNOLDS number - Abstract
The accurate estimation of the power number for closed clearance impellers holds significant importance in industries such as chemical, biochemical, paper and pulp, as well as paints, pigments, and polymers. Existing state‐of‐the‐art correlations for predicting power numbers, however, are inaccurate for impeller Reynolds number ReI>100. In this study, we compiled a dataset of 1470 data points from 15 research articles in the open literature, covering five types of impellers: (i) anchor; (ii) gate; (iii) single helical ribbon; (iv) double helical ribbon; and (v) helical ribbon with screw. Six machine learning models, namely artificial neural networks (ANN), CatBoost regressor, extra tree regressor, support vector regressor, random forest, and XGBoost regressor, were developed and compared. The results revealed that ANN emerged as the most efficient model, demonstrating the highest testing R2 value of 0.99 and the lowest testing MAPE of 7.3%. Further, we used the ANN model to develop a novel set of process correlations to estimate impeller power numbers for the industrially important anchor and double helical ribbon impellers: which significantly outperform the existing state‐of‐the‐art correlations available in literature. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Analysis of handmade paper by Raman spectroscopy combined with machine learning.
- Author
-
Yan, Chunsheng, Cheng, Zhongyi, Luo, Si, Huang, Chen, Han, Songtao, Han, Xiuli, Du, Yuandong, and Ying, Chaonan
- Subjects
- *
MACHINE learning , *RAMAN spectroscopy , *SUPPORT vector machines , *K-nearest neighbor classification , *PRINCIPAL components analysis , *RANDOM forest algorithms , *SPECTRAL imaging , *MULTISPECTRAL imaging - Abstract
Handmade paper is a major carrier and restoration material of traditional Chinese ancient books, calligraphies, and paintings. In this study, we carried out a Raman spectroscopy analysis of 18 types of handmade paper samples. The main components of the handmade paper were cellulose and lignin, according to the wavenumber and Raman vibration assignment. We divided its Raman spectrum into eight subbands. Five machine learning models were employed: principal component analysis (PCA), partial least squares (PLS), support vector machine (SVM), k‐nearest neighbors (KNN), and random forest (RF). The Raman spectral data were normalized, and the fluorescence envelope was subtracted using the airPLS algorithm to obtain four types of data, raw, normalized, defluorescence, and fluorescence data. An RF variable importance analysis of data processing showed that data normalization eliminated the intensity differences of fluorescence signals caused by lignin, which contained important information of raw materials and papermaking technology, let alone the data defluorescence. The data processing also reduced the importance of the average variables in almost all spectral bands. Nevertheless, the data processing is worthwhile because it significantly improves the accuracy of machine learning, and the information loss does not affect the prediction. Using the machine learning models of PCA, PLS, and SVM combined with linear regression (LR), KNN, and RF, the classification and prediction of handmade paper samples were realized. For almost all processed data, including the fluorescence data, PCA‐LR had the highest classification and prediction accuracy (R2 = 1) in almost all spectral bands. PLS‐LR and SVM‐LR had the second‐highest accuracies (R2 = 0.4–0.9), whereas KNN and RF had the lowest accuracies (R2 = 0.1–0.4) for full band spectral data. Our results suggest that the abundant information contained in Raman spectroscopy combined with powerful machine learning models could inspire further studies on handmade paper and related cultural relics. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
5. Content‐based and knowledge graph‐based paper recommendation: Exploring user preferences with the knowledge graphs for scientific paper recommendation.
- Author
-
Tang, Hao, Liu, Baisong, and Qian, Jiangbo
- Subjects
KNOWLEDGE graphs ,SCIENTIFIC knowledge ,DEEP learning ,CONVOLUTIONAL neural networks ,MACHINE learning ,RECOMMENDER systems ,USER-generated content - Abstract
Researchers usually face difficulties in finding scientific papers relevant to their research interests due to increasing growth. Recommender systems emerge as a leading solution to filter valuable items intelligently. Recently, deep learning algorithms, such as convolutional neural network, improved traditional recommendation technologies, for example, the graph‐based or content‐based methods. However, existing graph‐based methods ignore high‐order association between users and items on graphs, and content‐based methods ignore global features of texts for explicit user preferences. Therefore, this paper proposes a Content‐based and knowledge Graph‐based Paper Recommendation method (CGPRec), which uses a two‐layer self‐attention block to obtain global features of texts for more complete explicit user preferences, and proposes an improved graph convolutional network for modeling high‐order associations on the knowledge graph to mine implicit user preferences. And the knowledge graph in this paper is constructed with concept nodes, user nodes, paper nodes, and other meta‐data nodes. Experimental results on a public dataset, CiteULike‐a, and a real application log dataset, AHData, show that our model outperforms compared with baseline methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
6. Editorial: Best papers of the 14th International Conference on Software and System Processes (ICSSP 2020) and 15th International Conference on Global Software Engineering (ICGSE 2020).
- Author
-
Steinmacher, Igor, Clarke, Paul, Tuzun, Eray, and Britto, Ricardo
- Subjects
- *
SOFTWARE engineering , *SOFTWARE engineers , *SYSTEMS software , *CONFERENCES & conventions , *SOFTWARE maintenance , *COMPUTER software development - Abstract
Today's software industry is global, virtual, and depending more than ever on strong and reliable processes. Stakeholders and infrastructure are distributed across the globe, posing challenges that go beyond those with co‐located teams and servers. Software Engineering continues to be a complex undertaking, with projects challenged to meet expectations, especially regarding costs. We know that Software Engineering is an ever‐changing discipline, with the result that firms and their employees must regularly embrace new methods, tools, technologies, and processes. In 2020, the International Conference on Global Software Engineering (ICGSE) and the International Conference on Systems and Software Processes (ICSSP) joined forces aiming to create a holistic understanding of the software landscape both from the perspective of human and infrastructure distribution and also the processes to support software development. Unfortunately, these challenges have become even more personal to many more in 2020 due to the disruption introduced by the COVID‐19 pandemic, which forced both conferences to be held virtually. As an outcome of the joint event, we selected a set of the best papers from the two conferences, which were invited to submit extended versions to this Special Issue in the Journal of Software: Maintenance and Evolution. Dedicated committees were established to identify the best papers. Eight papers were invited and ultimately, seven of these invited papers have made it into this Special Issue. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
7. Spotlights are papers selected by editors published in peer‐reviewed journals that may be more regionally specific or appearing in languages other than English.
- Subjects
- *
ELECTRONIC journals , *MACHINE learning , *ENGLISH language , *HUMAN fingerprints - Abstract
This document highlights two studies published in the Asian Journal of Ecotoxicology. The first study focuses on the development of machine learning models to screen chemicals with hepatotoxicity, or liver toxicity. The models were trained using a dataset of 4014 chemicals and achieved good performance in predicting hepatotoxicity. The second study explores the use of machine learning methods to screen chemicals that induce autonomic dysfunction, a condition affecting the autonomic nervous system. The study developed a model using a dataset of 466 positive and 427 negative samples and identified structural alerts associated with autonomic dysfunction. Both studies provide valuable tools for screening and evaluating toxic chemicals. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
8. Evaluation method for moisture content of oil‐paper insulation based on segmented frequency domain spectroscopy: From curve fitting to machine learning.
- Author
-
Yao, Huanmin, Mu, Haibao, Ding, Ning, Zhang, Daning, Liang, ZhaoJie, Tian, Jie, and Zhang, Guanjun
- Subjects
- *
SPECTROMETRY , *MOISTURE , *MACHINE learning , *RANDOM forest algorithms , *DECISION trees - Abstract
In recent years, frequency domain spectroscopy (FDS) is often used to evaluate oil paper insulation state in power transformer bushing. But it is still very difficult to evaluate the moisture content accurately and quickly. In order to solve this problem, this paper proposes an intelligent algorithm based on random forest regression (RFR) to construct an efficient evaluation method through segmented FDS curves. Furthermore, the characteristics of FDS curves were studied and the intelligent method was compared with support vector regression (SVR) and deep neural networks (DNN). The results show that the dielectric loss, the real part and imaginary part of complex capacitance all move upward with the moisture increasing, so they can be used as the input feature of the evaluation model; The moisture content evaluation accuracy of the RFR model in the whole frequency band is higher than that of SVR and DNN models; With the increase of lower cut off frequency (FDS test stop frequency), the FDS test time is greatly shortened, and the accuracy of the RFR model can still meet the evaluation requirements. Therefore, the data in a compromise frequency band can be used to evaluate the moisture content of oil paper insulation accurately and quickly. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
9. Introduction to the virtual collection of papers on Artificial neural networks: applications in X‐ray photon science and crystallography.
- Author
-
Ekeberg, Tomas
- Subjects
- *
ARTIFICIAL neural networks , *DEEP learning , *CRYSTALLOGRAPHY , *ARTIFICIAL intelligence , *MACHINE learning , *PHOTONS - Abstract
Artificial intelligence is more present than ever, both in our society in general and in science. At the center of this development has been the concept of deep learning, the use of artificial neural networks that are many layers deep and can often reproduce human‐like behavior much better than other machine‐learning techniques. The articles in this collection are some recent examples of its application for X‐ray photon science and crystallography that have been published in Journal of Applied Crystallography. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. 3‐3: Invited Paper: Prediction Model for Visual Fatigue Caused by Smartphone Display Based on EEG Multi‐dimensional Features.
- Author
-
Shi, Yunyang, Tu, Yan, Wang, Lili, and Zhu, Nianfang
- Subjects
FEATURE extraction ,PREDICTION models ,ELECTROENCEPHALOGRAPHY ,SMARTPHONES - Abstract
In this study, a prediction model for visual fatigue is developed. As input, frequential and nonlinear features are extracted from multichannel EEG, and then dimensionally reduced. In the model, bidirectional LSTM and attention layers are combined for effective learning. As a result, 82.90% accuracy, 85.26% weighted precision, 82.90% weighted recall, and 84.02% weighted F1‐score were obtained. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Computational pathology definitions, best practices, and recommendations for regulatory guidance: a white paper from the Digital Pathology Association.
- Author
-
Abels, Esther, Pantanowitz, Liron, Aeffner, Famke, Zarella, Mark D, Laak, Jeroen, Bui, Marilyn M, Vemuri, Venkata NP, Parwani, Anil V, Gibbs, Jeff, Agosto‐Arroyo, Emmanuel, Beck, Andrew H, and Kozlowski, Cleopatra
- Subjects
ELECTRONIC paper ,BEST practices ,ARTIFICIAL neural networks ,PATHOLOGY - Abstract
In this white paper, experts from the Digital Pathology Association (DPA) define terminology and concepts in the emerging field of computational pathology, with a focus on its application to histology images analyzed together with their associated patient data to extract information. This review offers a historical perspective and describes the potential clinical benefits from research and applications in this field, as well as significant obstacles to adoption. Best practices for implementing computational pathology workflows are presented. These include infrastructure considerations, acquisition of training data, quality assessments, as well as regulatory, ethical, and cyber‐security concerns. Recommendations are provided for regulators, vendors, and computational pathology practitioners in order to facilitate progress in the field. © 2019 The Authors. The Journal of Pathology published by John Wiley & Sons Ltd on behalf of Pathological Society of Great Britain and Ireland. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
12. Distance versus Capillary Flow Dynamics‐Based Detection Methods on a Microfluidic Paper‐Based Analytical Device (μPAD).
- Author
-
Chung, Soo, Jennings, Christian M., and Yoon, Jeong‐Yeol
- Subjects
- *
CAPILLARY flow , *MICROFLUIDICS , *POINT-of-care testing , *CHEMICAL detectors , *DISTANCES - Abstract
In recent years, there has been high interest in paper‐based microfluidic sensors or microfluidic paper‐based analytical devices (μPADs) towards low‐cost, portable, and easy‐to‐use sensing for chemical and biological targets. μPAD allows spontaneous liquid flow without any external or internal pumping, as well as an innate filtration capability. Although both optical (colorimetric and fluorescent) and electrochemical detection have been demonstrated on μPADs, several limitations still remain, such as the need for additional equipment, vulnerability to ambient lighting perturbation, and inferior sensitivity. Herein, alternative detection methods on μPADs are reviewed to resolve these issues, including relatively well studied distance‐based measurements and the newer capillary flow dynamics‐based method. Detection principles, assay performance, strengths, and weaknesses are explained for these methods, along with their potential future applications towards point‐of‐care medical diagnostics and other field‐based applications. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
13. Does double‐blind peer review reduce bias? Evidence from a top computer science conference.
- Author
-
Sun, Mengyi, Barry Danfa, Jainabou, and Teplitskiy, Misha
- Subjects
PROFESSIONAL peer review ,COMPUTERS ,MANUSCRIPTS ,CONFERENCES & conventions ,MACHINE learning ,MANN Whitney U Test ,CITATION analysis ,INFORMATION science ,RESEARCH bias - Abstract
Peer review is essential for advancing scientific research, but there are long‐standing concerns that authors' prestige or other characteristics can bias reviewers. Double‐blind peer review has been proposed as a way to reduce reviewer bias, but the evidence for its effectiveness is limited and mixed. Here, we examine the effects of double‐blind peer review by analyzing the review files of 5,027 papers submitted to a top computer science conference that changed its reviewing format from single‐ to double‐blind in 2018. First, we find that the scores given to the most prestigious authors significantly decreased after switching to double‐blind review. However, because many of these papers were above the threshold for acceptance, the change did not affect paper acceptance significantly. Second, the inter‐reviewer disagreement increased significantly in the double‐blind format. Third, papers rejected in the single‐blind format are cited more than those rejected under double‐blind, suggesting that double‐blind review better excludes poorer quality papers. Lastly, an apparently unrelated change in the rating scale from 10 to 4 points likely reduced prestige bias significantly such that papers' acceptance was affected. These results support the effectiveness of double‐blind review in reducing biases, while opening new research directions on the impact of peer‐review formats. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
14. 85‐1: Invited Paper: A Novel OLED Material Discovery based on AI Technology.
- Author
-
Kim, Hoilim, Kim, Seran, Yoo, Dongsun, Kim, Gyeounghun, Koh, Eunkyung, Kim, Jihye, Park, Saerom, Kim, Sohae, Shin, Hyosup, Cho, Hyunguk, and Baek, Seungin
- Subjects
DELAYED fluorescence ,MACHINE learning ,ARTIFICIAL intelligence ,LIGHT emitting diodes ,MOLECULAR dynamics - Abstract
We report a novel OLED material discovery process and several applications based on AI technology. This process in which six AI modules that generate molecular structures with active learning algorithm, predict multiple properties, analyze novelty, predict synthetic scheme, predict relative synthesizability and predict device characteristics are linked one after another. Also, we introduce some cases in which materials designed by this process were actually synthesized and applied to devices for evaluation to confirm the improvement of characteristics. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. P‐103: Multi Path Control System for AGV Based on Digital Twin in Display Module Process.
- Author
-
Kim, Yonggu, Lee, Seungjoo, and Roh, Cheollae
- Subjects
DIGITAL twins ,ELECTRONIC paper ,TRAFFIC congestion ,MACHINE learning ,INTELLIGENT transportation systems ,AUTOMATED guided vehicle systems - Abstract
Increasing display process complexity, Automated Guided Vehicle becomes more important and the delivery time constraint is getting shorter. The complicated process causes traffic congestion and deadlocks between AGVs. Preventing these issues, this paper suggests digital twin technology that allows AGV to choose the best route among various paths. This research validates the AGV simulation for the display plant. The result shows AGV average delivery time was shortened by 14.1% compared to the traditional algorithm. In addition, the AGV capacity rate can be reduced by 21.5% and the number of AGVs can be reduced by 10.1%. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
16. How to cheat the page limit.
- Author
-
Duivesteijn, Wouter, Hess, Sibylle, and Du, Xin
- Subjects
DATA mining ,DATA science ,MACHINE learning ,SWINDLERS & swindling - Abstract
Every conference imposing a limit on the length of submissions must deal with the problem of page limit cheating: authors tweaking the parameters of the game such that they can squeeze more content into their paper. We claim that this problem is endemic, although we lack the data to formally prove this. Instead, this paper provides a far from exhaustive summary of ways to cheat the page limit, a case study involving the papers accepted for the Research and Applied Data Science tracks at the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECMLPKDD) 2019, and a discussion of ways for program chairs to tackle this problem. Of the 130 accepted papers in these two ECMLPKDD 2019 tracks, 68 satisfied the page limit; 62 (47.7%) turned out to spill over the page limit, by up to as much as 50%. To misappropriate a phrase from Darrell Huff's "How to Lie with Statistics," we intend for this paper not to be a manual for swindlers; instead, nefarious paper authors already know these tricks, and honest program chairs must learn them in self‐defense. This article is categorized under:Commercial, Legal, and Ethical Issues > Fairness in Data Mining [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
17. Guest Editorial: Artificial intelligence‐empowered reliable forecasting for energy sectors.
- Author
-
Mahmoud, Karar, Guerrero, Josep M., Abdel‐Nasser, Mohamed, and Yorino, Naoto
- Subjects
ENERGY industries ,ARTIFICIAL neural networks ,MACHINE learning ,FORECASTING ,QUANTILE regression ,CONVOLUTIONAL neural networks ,DEMAND forecasting - Abstract
This document is a guest editorial from the journal IET Generation, Transmission & Distribution. It discusses the use of artificial intelligence (AI) in reliable forecasting for energy sectors. The editorial highlights the challenges of integrating renewable energy sources and fluctuating electricity demand, and emphasizes the importance of accurate forecasting for system operators. The document also provides summaries of several papers included in a special issue on AI-empowered forecasting in energy sectors, covering topics such as load forecasting, wind power prediction, and control parameter optimization. The editorial concludes by recommending further research and practical implementations of AI approaches in the energy sectors. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
18. 72‐4: Invited Paper: Synthetic Defect Generation for Display Front‐of‐Screen Quality Inspection: A Survey.
- Author
-
Mou, Shancong, Cao, Meng, Hong, Zhendong, Huang, Ping, Shan, Jiulong, and Shi, Jianjun
- Subjects
MASS production ,MACHINE learning ,DEEP learning ,MANUFACTURING processes ,EVALUATION methodology - Abstract
Display front‐of‐screen (FOS) quality inspection is essential for the mass production of displays in the manufacturing process. However, the severe imbalanced data, especially the limited number of defective samples, has been a long‐standing problem that hinders the successful application of deep learning algorithms. Synthetic defect data generation can help address this issue. This paper reviews the state‐of‐the‐art synthetic data generation methods and the evaluation metrics that can potentially be applied to display FOS quality inspection tasks. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
19. 37‐2: Invited Paper: Enabling Augmented‐Reality Near‐Eye and Head‐Up Displays with Neural Holography.
- Author
-
Choi, Suyeon, Peng, Yifan, Gopakumar, Manu, Kim, Jonghyun, and Wetzstein, Gordon
- Subjects
HOLOGRAPHY ,HOLOGRAPHIC displays ,HEAD-up displays ,DISPLAY systems ,AUGMENTED reality ,HEAD-mounted displays ,ARTIFICIAL intelligence - Abstract
Holography promises unprecedented capabilities for augmented reality (AR) display systems, but the current achievable image quality is quite limited. We leverage Neural Holography, a family of artificial intelligence‐driven computer‐generated holography (CGH) techniques, to enable two holographic see‐through AR displays: a lightguide coupled near‐eye display (NED) and a partial reflective head‐up display (HUD). Experimental results demonstrate the promise of enabling next‐generation AR 3D holographic displays. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
20. 31‐2: Student Paper: Fermi Level Prediction of Solution‐processed Ultra‐wide Band gap a‐Ga2Ox via Supervised Machine Learning Models.
- Author
-
Purnawati, Diki, Regonia, Paul Rossener, Bermundo, Juan Paolo, Ikeda, Kazushi, and Uraoka, Yukiharu
- Subjects
SUPERVISED learning ,FERMI level ,BAND gaps ,MACHINE learning ,RANDOM forest algorithms ,ULTRA-wideband radar - Abstract
This work presents machine learning (ML) assisted Fermi level prediction of solution‐processed ultra‐wide bandgap (UWB) amorphous gallium oxide (a‐Ga2Ox) which can significantly accelerate the fabrication of semiconducting UWB a‐Ga2Ox‐based material for future display application. Different models such as Kernel Ridge Regression (KRR), Support Vector Regression (SVR) and Random Forest Regression (RFR) were trained with empirical features, including experimental thickness, annealing temperature and environment during the solution‐processed UWB a‐Ga2Ox film fabrication. This work is a big step towards rapid and cost‐effective optimization method of fabricating UWB a‐Ga2Ox‐based devices. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
21. Mining Information from Collections of Papers: Illustrative Analysis of Groundwater and Disease.
- Author
-
Zhang, Yiding, Ji, Xiaonan, Ibaraki, Motomu, and Schwartz, Franklin W.
- Subjects
- *
DATA mining , *MACHINE learning , *GROUNDWATER & health , *WATERBORNE infection , *GROUNDWATER pollution - Abstract
The academic world is driven by scholarly research and publications. Yet, for many fields, the volume of published research and the associated knowledge base have been expanding exponentially for decades. The result is that scientists are literally drowning in data and information. There are strategies and approaches that could help with this problem. The goal of this paper is to demonstrate the power of computer‐based approaches such as data mining and machine learning to evaluate large collections of papers. The objective is to conduct a systematic analysis of research related to the emerging area of groundwater‐related diseases. More specifically, the analysis of information from the database of papers will examine systematics in the research topics, the inter‐relationships among multiple diseases, contaminants, and groundwater, and discover styles of research associated with groundwater and disease. The analysis uses 426 papers (1971 to 2017) retrieved from a MEDLINE bibliographic database, PubMed, given the search terms "groundwater" and "disease." We developed tools that take care of necessary text processing steps, which lead naturally to clustering and visualization techniques that demonstrate published research. The resulting 2D article map shows how the collection of papers is subdivided into 11 article clusters. The cluster topics were determined by analyzing keywords or common words contained in the articles' titles, abstracts, and key words. We found that research on water‐related disease in groundwater primarily focuses on two types of contaminants—chemical compounds and pathogens. Cancer and diarrhea are two major diseases associated with groundwater contamination. According to the systematic analysis, the study of this area is still growing. Article impact statement: Using data mining and machine learning techniques to study the research strands of groundwater‐related disease from the database. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
22. Research on case preprocessing based on deep learning.
- Author
-
Zhang, Chuyue, Cai, Manchun, Zhao, Xiaofan, and Wang, Dawei
- Subjects
DEEP learning ,MACHINE learning ,DATA mining ,CONFERENCE papers ,DATA quality - Abstract
Considering the problem of missing fields in the criminal case system, this article proposes a deep learning algorithm to extract the features of the case description and fill in the missing value. Due to Chinese expressions and characteristics of criminal cases, we make both character vectors and word vectors to present text embedding. Character vectors are from bert model. Word vector is trained by long short‐term memory model with attention. The experiment uses 13,890 data totally. This work is an extension of our short conference proceeding paper. The results show that the combination of characters and words can effectively improve the accuracy of the conference paper by 9%. This is the first time to cascade the character and word dimensions on the criminal case information preprocess and it can provide higher quality data especially for the crime data mining. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
23. Machine learning and big data analytics in bipolar disorder: A position paper from the International Society for Bipolar Disorders Big Data Task Force.
- Author
-
Passos, Ives C., Ballester, Pedro L., Barros, Rodrigo C., Librenza‐Garcia, Diego, Mwangi, Benson, Birmaher, Boris, Brietzke, Elisa, Hajek, Tomas, Lopez Jaramillo, Carlos, Mansur, Rodrigo B., Alda, Martin, Haarman, Bartholomeus C. M., Isometsa, Erkki, Lam, Raymond W., McIntyre, Roger S., Minuzzi, Luciano, Kessing, Lars V., Yatham, Lakshmi N., Duffy, Anne, and Kapczinski, Flavio
- Subjects
- *
BIG data , *BIPOLAR disorder , *TASK forces , *MACHINE learning , *SCIENTIFIC literature - Abstract
Objectives: The International Society for Bipolar Disorders Big Data Task Force assembled leading researchers in the field of bipolar disorder (BD), machine learning, and big data with extensive experience to evaluate the rationale of machine learning and big data analytics strategies for BD. Method: A task force was convened to examine and integrate findings from the scientific literature related to machine learning and big data based studies to clarify terminology and to describe challenges and potential applications in the field of BD. We also systematically searched PubMed, Embase, and Web of Science for articles published up to January 2019 that used machine learning in BD. Results: The results suggested that big data analytics has the potential to provide risk calculators to aid in treatment decisions and predict clinical prognosis, including suicidality, for individual patients. This approach can advance diagnosis by enabling discovery of more relevant data‐driven phenotypes, as well as by predicting transition to the disorder in high‐risk unaffected subjects. We also discuss the most frequent challenges that big data analytics applications can face, such as heterogeneity, lack of external validation and replication of some studies, cost and non‐stationary distribution of the data, and lack of appropriate funding. Conclusion: Machine learning‐based studies, including atheoretical data‐driven big data approaches, provide an opportunity to more accurately detect those who are at risk, parse‐relevant phenotypes as well as inform treatment selection and prognosis. However, several methodological challenges need to be addressed in order to translate research findings to clinical settings. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
24. Introduction to the special issue on 'Advanced techniques, methods and applications for an integrated approach to the geophysical prospecting'.
- Author
-
Persico, Raffaele
- Subjects
GEOPHYSICAL prospecting ,MACHINE learning ,GROUND penetrating radar ,CULTURAL property ,GEOPHYSICS - Abstract
This document is an introduction to a special issue of the journal Geophysical Prospecting on "Advanced techniques, methods, and applications for an integrated approach to geophysical prospecting." The issue focuses on the use of integrated approaches in geophysics, which involve the intelligent use of multiple techniques and the combination of their results. The special issue includes 14 papers that cover a range of techniques such as seismic, geoelectric, ground-penetrating radar (GPR), and magneto-telluric, as well as various applications including mining, cultural heritage, and geological evaluations. The papers explore topics such as reservoir modeling, water reservoir mapping, machine learning algorithms, mine structure detection, and archaeological investigations. The document concludes with acknowledgments and a statement that the data are available upon request. [Extracted from the article]
- Published
- 2023
- Full Text
- View/download PDF
25. A systematic review of Green AI.
- Author
-
Verdecchia, Roberto, Sallou, June, and Cruz, Luís
- Subjects
ARTIFICIAL intelligence ,CARBON emissions ,CLEAN energy ,SUSTAINABILITY ,ECOLOGICAL impact ,MACHINE learning - Abstract
With the ever‐growing adoption of artificial intelligence (AI)‐based systems, the carbon footprint of AI is no longer negligible. AI researchers and practitioners are therefore urged to hold themselves accountable for the carbon emissions of the AI models they design and use. This led in recent years to the appearance of researches tackling AI environmental sustainability, a field referred to as Green AI. Despite the rapid growth of interest in the topic, a comprehensive overview of Green AI research is to date still missing. To address this gap, in this article, we present a systematic review of the Green AI literature. From the analysis of 98 primary studies, different patterns emerge. The topic experienced a considerable growth from 2020 onward. Most studies consider monitoring AI model footprint, tuning hyperparameters to improve model sustainability, or benchmarking models. A mix of position papers, observational studies, and solution papers are present. Most papers focus on the training phase, are algorithm‐agnostic or study neural networks, and use image data. Laboratory experiments are the most common research strategy. Reported Green AI energy savings go up to 115%, with savings over 50% being rather common. Industrial parties are involved in Green AI studies, albeit most target academic readers. Green AI tool provisioning is scarce. As a conclusion, the Green AI research field results to have reached a considerable level of maturity. Therefore, from this review emerges that the time is suitable to adopt other Green AI research strategies, and port the numerous promising academic results to industrial practice. This article is categorized under:Technologies > Machine Learning [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. Three machine learning algorithms and their utility in exploring risk factors associated with primary cesarean section in low‐risk women: A methods paper.
- Author
-
Clark, Rebecca R. S. and Hou, Jintong
- Subjects
OXYTOCIN ,HOSPITALS ,RESEARCH evaluation ,MACHINE learning ,REGRESSION analysis ,PREGNANT women ,RISK assessment ,PREGNANCY outcomes ,RESEARCH funding ,CESAREAN section ,DATA analysis software ,OBESITY in women ,ALGORITHMS ,SECONDARY analysis ,PROBABILITY theory ,DISEASE complications - Abstract
Machine learning, a branch of artificial intelligence, is increasingly used in health research, including nursing and maternal outcomes research. Machine learning algorithms are complex and involve statistics and terminology that are not common in health research. The purpose of this methods paper is to describe three machine learning algorithms in detail and provide an example of their use in maternal outcomes research. The three algorithms, classification and regression trees, least absolute shrinkage and selection operator, and random forest, may be used to understand risk groups, select variables for a model, and rank variables' contribution to an outcome, respectively. While machine learning has plenty to contribute to health research, it also has some drawbacks, and these are discussed as well. To provide an example of the different algorithms' function, they were used on a completed cross‐sectional study examining the association of oxytocin total dose exposure with primary cesarean section. The results of the algorithms are compared to what was done or found using more traditional methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
27. Trends in quantum reinforcement learning: State‐of‐the‐arts and the road ahead.
- Author
-
Park, Soohyun and Kim, Joongheon
- Subjects
FEDERATED learning ,REINFORCEMENT learning ,MACHINE learning ,QUANTUM computing ,COMPUTER software testing ,DEEP learning - Abstract
This paper presents the basic quantum reinforcement learning theory and its applications to various engineering problems. With the advances in quantum computing and deep learning technologies, various research works have focused on quantum deep learning and quantum machine learning. In this paper, quantum neural network (QNN)‐based reinforcement learning (RL) models are discussed and introduced. Moreover, the pros of the QNN‐based RL algorithms and models, such as fast training, high scalability, and efficient learning parameter utilization, are presented along with various research results. In addition, one of the well‐known multi‐agent extensions of QNN‐based RL models, the quantum centralized‐critic and multiple‐actor network, is also discussed and its applications to multi‐agent cooperation and coordination are introduced. Finally, the applications and future research directions are introduced and discussed in terms of federated learning, split learning, autonomous control, and quantum deep learning software testing. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Exploring the opportunity of using machine learning to support the system dynamics method: Comment on the paper by Edali and Yücel.
- Author
-
Duggan, Jim
- Subjects
ALGORITHMS ,COMPUTER simulation ,DECISION making ,MACHINE learning ,HUMAN services programs - Abstract
The author presents comments on a paper on the use of machine learning to support the system dynamics method. Topics discussed include its interpretation of simulation models and explanation of policy analysis, and the emerging view whereby dynamic problems from endogenous feedback structures can be tackled via wider tools and methodological approaches. Also noted is the resulting potential for greater insights into the modelling process.
- Published
- 2020
- Full Text
- View/download PDF
29. 60‐3: Invited Paper: Machine Learning Approaches to Active Stylus for Capacitive Touch Screen Panel Applications.
- Author
-
Nam, Hyoungsik, Seol, Ki-Hyuk, and Park, Seungjun
- Subjects
MACHINE learning ,SUPPORT vector machines ,ANOMALY detection (Computer security) ,TOUCH screens ,CLASSIFICATION algorithms - Abstract
This paper introduces machine learning approaches on adding the stylus‐touch to the capacitive touch screen technology. The proposed schemes can discriminate the stylus‐touch from finger‐touch as well as no‐touch by means of classification algorithms using support vector machine and anomaly detection. The high frequency pulses are sent from a stylus to a touch screen and the receiver classifies the received sample sequences into three classes of no‐touch, finger‐touch, and stylus‐touch. In addition, some possible applications of data transmission and user authentication are demonstrated. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
30. Special issue on the best papers of the Conference on Intelligent Data Understanding (CIDU 2010).
- Author
-
Srivastava, Ashok N. and Chawla, Nitesh V.
- Subjects
CONFERENCES & conventions ,DATA mining ,MACHINE learning ,CLIMATOLOGY ,ASTRONOMY ,AERONAUTICAL safety measures - Abstract
The article discusses the highlights of the Conference on Intelligent Data Understanding held at the Computer History Museum in Mountain View, California on October 5-6, 2010. Topics included the implementation of the methods attained in data mining and machine learning on issues dealing with earth sciences, space sciences and systems health management. Problems dealing with changes in the climate and environment, astronomical data flux and safety in aviation were also tackled.
- Published
- 2011
- Full Text
- View/download PDF
31. 16‐4: Invited Paper: Region‐Based Machine Learning for OLED Mura Defects Detection.
- Author
-
Lee, Janghwan
- Subjects
MACHINE learning ,AUTOMATIC optical inspection ,OPTICAL images ,ORGANIC light emitting diodes ,DIGITAL cameras - Abstract
Automatic inspection of Mura defects is a challenging task in display manufacturing. Due to the nature of Mura defects, which appear as brightness variances in the low‐contrast images captured by the optical inspection camera, the defects are extremely difficult to detect because they show no clear edges from their surroundings while the image backgrounds usually present uneven illumination. In this paper, I propose an effective way to detect two types of Mura defects using a region‐based machine learning approach. My research includes three components: 1) creating a quality dataset from the raw optical inspection images, 2) designing a region‐based machine learning model with a preprocessor, a candidate detector, a feature extractor, and a classifier, and 3) an adversarial training and evaluation method to overcome the inconsistent data labels. Applying real panel images from the display manufacturing line as my test set, my trained model achieved a recall rate of 100% and a precision rate of over 90%. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
32. Call for Papers: Issues and Practice in Applying Machine Learning in Educational Measurement.
- Subjects
- *
MACHINE learning , *EDUCATIONAL tests & measurements , *ARTIFICIAL intelligence - Abstract
For example, a manuscript that examines the application of machine learning methods in reviewing irregularity reports from test administration is in scope. Machine learning, or artificial intelligence as a broader term, has been very popular in recent years. [Extracted from the article]
- Published
- 2022
- Full Text
- View/download PDF
33. Learning‐driven ubiquitous mobile edge computing: Network management challenges for future generation Internet of Things.
- Author
-
Donta, Praveen Kumar, Monteiro, Edmundo, Dehury, Chinmaya Kumar, and Murturi, Ilir
- Subjects
MOBILE computing ,MOBILE learning ,EDGE computing ,INTERNET of things ,ARTIFICIAL intelligence ,AMBIENT intelligence - Abstract
Learning-driven ubiquitous mobile edge computing: Network management challenges for future generation Internet of Things Keywords: artificial intelligence; Internet of Things; machine learning; mobile edge computing; network Management EN artificial intelligence Internet of Things machine learning mobile edge computing network Management 1 4 4 09/13/23 20230901 NES 230901 Ubiquitous edge computing facilitates efficient cloud services near mobile devices, enabling mobile edge computing (MEC) to offer services more efficiently by presenting storage and processing capability within the proximity of mobile devices and in general IoT domains. Fifth, I Jafar et al. i proposes "A blockchain-enabled security management framework for mobile edge computing", which guarantees secure data storage and includes blockchain features like immutability and traceability. [Extracted from the article]
- Published
- 2023
- Full Text
- View/download PDF
34. How Interpretable Machine Learning Can Benefit Process Understanding in the Geosciences.
- Author
-
Jiang, Shijie, Sweet, Lily‐belle, Blougouras, Georgios, Brenning, Alexander, Li, Wantong, Reichstein, Markus, Denzler, Joachim, Shangguan, Wei, Yu, Guo, Huang, Feini, and Zscheischler, Jakob
- Subjects
MACHINE learning ,ARTIFICIAL intelligence ,EARTH currents ,ARTIFICIAL languages ,RESEARCH questions - Abstract
Interpretable Machine Learning (IML) has rapidly advanced in recent years, offering new opportunities to improve our understanding of the complex Earth system. IML goes beyond conventional machine learning by not only making predictions but also seeking to elucidate the reasoning behind those predictions. The combination of predictive power and enhanced transparency makes IML a promising approach for uncovering relationships in data that may be overlooked by traditional analysis. Despite its potential, the broader implications for the field have yet to be fully appreciated. Meanwhile, the rapid proliferation of IML, still in its early stages, has been accompanied by instances of careless application. In response to these challenges, this paper focuses on how IML can effectively and appropriately aid geoscientists in advancing process understanding—areas that are often underexplored in more technical discussions of IML. Specifically, we identify pragmatic application scenarios for IML in typical geoscientific studies, such as quantifying relationships in specific contexts, generating hypotheses about potential mechanisms, and evaluating process‐based models. Moreover, we present a general and practical workflow for using IML to address specific research questions. In particular, we identify several critical and common pitfalls in the use of IML that can lead to misleading conclusions, and propose corresponding good practices. Our goal is to facilitate a broader, yet more careful and thoughtful integration of IML into Earth science research, positioning it as a valuable data science tool capable of enhancing our current understanding of the Earth system. Plain Language Summary: Artificial Intelligence is a rapidly advancing field, in which Interpretable Machine Learning (IML) is seen as having the potential to significantly improve our understanding of Earth's complex environmental systems. IML goes beyond the predictive power of machine learning models, focusing instead on uncovering the relationships within the data that are revealed by the model's learning process. However, there is still a lack of straightforward, practical domain‐specific guidelines for geoscientists that facilitate both broader and more careful application in the field. In this paper, we aim to demonstrate the real‐world benefits of IML in typical geoscientific analysis. We provide a clear, step‐by‐step workflow that shows how IML can be used to address specific questions. We also point out some common pitfalls in using IML and offer solutions to avoid them. Our goal is to make IML more accessible and useful to a wider range of geoscientists, and we believe that IML, if used properly and thoughtfully, can become an essential and valuable tool to advance our understanding of complex Earth systems. Key Points: We demonstrate the broader relevance of Interpretable Machine Learning (IML) to most geoscientists and underexplored opportunities for its useWe describe a workflow for the effective use of IML while cautioning against potential and common pitfallsWe suggest good practices for its adoption and advocate for more careful application to ensure reliable and robust insights for the field [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. A review of asset management using artificial intelligence‐based machine learning models: Applications for the electric power and energy system.
- Author
-
Rajora, Gopal Lal, Sanz‐Bobi, Miguel A., Tjernberg, Lina Bertling, and Urrea Cabus, José Eduardo
- Subjects
ARTIFICIAL intelligence ,ASSET management ,ASSET protection ,MACHINE learning ,DEEP learning ,SUSTAINABILITY - Abstract
Power system protection and asset management present persistent technical challenges, particularly in the context of the smart grid and renewable energy sectors. This paper aims to address these challenges by providing a comprehensive assessment of machine learning applications for effective asset management in power systems. The study focuses on the increasing demand for energy production while maintaining environmental sustainability and efficiency. By harnessing the power of modern technologies such as artificial intelligence (AI), machine learning (ML), and deep learning (DL), this research explores how ML techniques can be leveraged as powerful tools for the power industry. By showcasing practical applications and success stories, this paper demonstrates the growing acceptance of machine learning as a significant technology for current and future business needs in the power sector. Additionally, the study examines the barriers and difficulties of large‐scale ML deployment in practical settings while exploring potential opportunities for these tactics. Through this overview, insights into the transformative potential of ML in shaping the future of power system asset management are provided. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. Nowcasting Earthquakes With Stochastic Simulations: Information Entropy of Earthquake Catalogs.
- Author
-
Rundle, John B., Baughman, Ian, and Zhang, Tianjian
- Subjects
EARTHQUAKES ,EARTHQUAKE aftershocks ,ENTROPY (Information theory) ,MACHINE learning ,EARTHQUAKE hazard analysis ,RECEIVER operating characteristic curves ,CATALOGS ,ENTROPY - Abstract
Earthquake nowcasting has been proposed as a means of tracking the change in large earthquake potential in a seismically active area. The method was developed using observable seismic data, in which probabilities of future large earthquakes can be computed using Receiver Operating Characteristic methods. Furthermore, analysis of the Shannon information content of the earthquake catalogs has been used to show that there is information contained in the catalogs, and that it can vary in time. So an important question remains, where does the information originate? In this paper, we examine this question using stochastic simulations of earthquake catalogs. Our catalog simulations are computed using an Earthquake Rescaled Aftershock Seismicity ("ERAS") stochastic model. This model is similar in many ways to other stochastic seismicity simulations, but has the advantage that the model has only 2 free parameters to be set, one for the aftershock (Omori‐Utsu) time decay, and one for the aftershock spatial migration away from the epicenter. Generating a simulation catalog and fitting the two parameters to the observed catalog such as California takes only a few minutes of wall clock time. While clustering can arise from random, Poisson statistics, we show that significant information in the simulation catalogs arises from the "non‐Poisson" power‐law aftershock clustering, implying that the practice of de‐clustering observed catalogs may remove information that would otherwise be useful in forecasting and nowcasting. We also show that the nowcasting method provides similar results with the ERAS model as it does with observed seismicity. Plain Language Summary: Earthquake nowcasting was proposed as a means of tracking the change in the potential for large earthquakes in a seismically active area, using the record of small earthquakes. The method was developed using observed seismic data, in which probabilities of future large earthquakes can be computed using machine learning methods that were originally developed with the advent of radar in the 1940s. These methods are now being used in the development of machine learning and artificial intelligence models in a variety of applications. In recent times, methods to simulate earthquakes using the observed statistical laws of earthquake seismicity have been developed. One of the advantages of these stochastic models is that it can be used to analyze the various assumptions that are inherent in the analysis of seismic catalogs of earthquakes. In this paper, we analyze the importance of the space‐time clustering that is often observed in earthquake seismicity. We find that the clustering is the origin of information that makes the earthquake nowcasting methods possible. We also find that a common practice of "aftershock de‐clustering", often used in the analysis of these catalogs, removes information about future large earthquakes. Key Points: Earthquake nowcasting tracks the change in the potential for large earthquakes, using information contained in seismic catalogsWe analyze the information contained in the space‐time clustering that is observed in earthquake seismicityWe find that "aftershock de‐clustering" of catalogs removes information about future large earthquakes that the nowcasting method uses [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Predicting coauthorship using bibliographic network embedding.
- Author
-
Zhu, Yongjun, Quan, Lihong, Chen, Pei‐Ying, Kim, Meen Chul, and Che, Chao
- Subjects
DECISION trees ,BIBLIOGRAPHY ,COMPUTER science ,MACHINE learning ,BIBLIOGRAPHICAL citations ,INFORMATION science ,INTERPROFESSIONAL relations ,RESEARCH funding ,PREDICTION models ,LOGISTIC regression analysis ,AUTHORSHIP - Abstract
Coauthorship prediction applies predictive analytics to bibliographic data to predict authors who are highly likely to be coauthors. In this study, we propose an approach for coauthorship prediction based on bibliographic network embedding through a graph‐based bibliographic data model that can be used to model common bibliographic data, including papers, terms, sources, authors, departments, research interests, universities, and countries. A real‐world dataset released by AMiner that includes more than 2 million papers, 8 million citations, and 1.7 million authors were integrated into a large bibliographic network using the proposed bibliographic data model. Translation‐based methods were applied to the entities and relationships to generate their low‐dimensional embeddings while preserving their connectivity information in the original bibliographic network. We applied machine learning algorithms to embeddings that represent the coauthorship relationships of the two authors and achieved high prediction results. The reference model, which is the combination of a network embedding size of 100, the most basic translation‐based method, and a gradient boosting method achieved an F1 score of 0.9 and even higher scores are obtainable with different embedding sizes and more advanced embedding methods. Thus, the strengths of the proposed approach lie in its customizable components under a unified framework. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. 81‐2: Invited Paper: Neural Network Based Quantitative Evaluation of Display Non‐Uniformity Corresponds Well with Human Visual Evaluation.
- Author
-
Tsutsukawa, Kazuki, Kobayashi, Manabu, and Bamba, Yusuke
- Subjects
PEARSON correlation (Statistics) - Abstract
We developed a neural network‐based method for evaluation of display luminance and color non‐uniformity (which we call Mura). We studied a correlation between our developed method and human visual evaluation because visual evaluation is the gold standard for Mura evaluation. We achieved Pearson correlation coefficient of 0.82. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
39. 37‐1: Invited Paper: 3D Computer Vision Based on Machine Learning with Deep Neural Networks.
- Author
-
Vodrahalli, Kailas and Bhowmik, Achintya K.
- Subjects
COMPUTER vision ,THREE-dimensional display systems ,ARTIFICIAL neural networks - Abstract
Recent advances in the field of computer vision can be attributed to the emergence of deep learning techniques, in particular convolutional neural networks. Neural networks, partially inspired by the brain’s visual cortex, enable a computer to “learn” the most important features of the images it is shown in relation to a specific, specified task. Given sufficient data and time, (deep) convolutional neural networks offer more easily designed, more generalizable, and significantly more accurate end‐to‐end systems than is possible with previously employed computer vision techniques. This review paper seeks to provide an overview of deep learning in the field of computer vision with an emphasis on recent progress in tasks involving 3D visual data. Through a backdrop of the mammalian visual processing system, we also hope to provide inspiration for future advances in automated visual processing. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
40. P‐17: All Weather‐Robust Image Quality Enhancement based on Image Feature Fusion and Multiscale Degradation Profile.
- Author
-
Ryu, Seungchul, Yoo, Hyunjin, and Akhavan, Tara
- Subjects
IMAGE intensifiers ,IMAGE enhancement (Imaging systems) ,IMAGE fusion ,SEVERE storms - Abstract
Severe weather conditions often induce degraded camera output images in automotive applications, which decreases the visibility of drivers. In order to address this problem, this paper proposes a multi‐scale degradation profile‐based image enhancement framework. The proposed framework is composed of a feature fusion module, a multi‐scale degradation profile module, and an image enhancement module. The qualitative evaluation, ablation study, and quantitative evaluation proved the advantages and practicability of the proposed framework. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. The most efficient machine learning algorithms in stroke prediction: A systematic review.
- Author
-
Asadi, Farkhondeh, Rahimi, Milad, Daeechini, Amir Hossein, and Paghe, Atefeh
- Subjects
MACHINE learning ,ARTIFICIAL intelligence ,SUPPORT vector machines ,RANDOM forest algorithms ,CAUSES of death - Abstract
Abstrac: Background and Aims: Stroke is one of the most common causes of death worldwide, leading to numerous complications and significantly diminishing the quality of life for those affected. The purpose of this study is to systematically review published papers on stroke prediction using machine learning algorithms and introduce the most efficient machine learning algorithms and compare their performance. The papers have published in period from 2019 to August 2023. Methods: The authors conducted a systematic search in PubMed, Scopus, Web of Science, and IEEE using the keywords "Artificial Intelligence," "Predictive Modeling," "Machine Learning," "Stroke," and "Cerebrovascular Accident" from 2019 to August 2023. Results: Twenty articles were included based on the inclusion criteria. The Random Forest (RF) algorithm was introduced as the best and most efficient stroke ML algorithm in 25% of the articles (n = 5). In addition, in other articles, Support Vector Machines (SVM), Stacking and XGBOOST, DSGD, COX& GBT, ANN, NB, and RXLM algorithms were introduced as the best and most efficient ML algorithms in stroke prediction. Conclusion: This research has shown a rapid increase in using ML algorithms to predict stroke, with significant improvements in model accuracy in recent years. However, no model has reached 100% accuracy or is entirely error‐free. Variations in algorithm efficiency and accuracy stem from differences in sample sizes, datasets, and data types. Further studies should focus on consistent datasets, sample sizes, and data types for more reliable outcomes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. Utilizing patient data: A tutorial on predicting second cancer with machine learning models.
- Author
-
Sadeghi, Hossein, Seif, Fatemeh, Farahani, Erfan Hatamabadi, Khanmohammadi, Soraya, and Nahidinezhad, Shahla
- Subjects
MACHINE learning ,RADIATION doses ,DECISION trees ,INDIVIDUALIZED medicine ,RADIOTHERAPY - Abstract
Background: The article explores the potential risk of secondary cancer (SC) due to radiation therapy (RT) and highlights the necessity for new modeling techniques to mitigate this risk. Methods: By employing machine learning (ML) models, specifically decision trees, in the research process, a practical framework is established for forecasting the occurrence of SC using patient data. Results & Discussion: This framework aids in categorizing patients into high‐risk or low‐risk groups, thereby enabling personalized treatment plans and interventions. The paper also underscores the many factors that contribute to the likelihood of SC, such as radiation dosage, patient age, and genetic predisposition, while emphasizing the limitations of current models in encompassing all relevant parameters. These limitations arise from the non‐linear dependencies between variables and the failure to consider factors such as genetics, hormones, lifestyle, radiation from secondary particles, and imaging dosage. To instruct and assess ML models for predicting the occurrence of SC based on patient data, the paper utilizes a dataset consisting of instances and attributes. Conclusion: The practical implications of this research lie in enhancing our understanding and prediction of SC following RT, facilitating personalized treatment approaches, and establishing a framework for leveraging patient data within the realm of ML models. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
43. Electrical line fault prediction using a novel grey wolf optimization algorithm based on multilayer perceptron.
- Author
-
Zhang, Yufei
- Abstract
Grey wolf optimization algorithm (GWO) has achieved great results in the optimization of neural network parameters. However, it has some problems such as insufficient precision, poor robustness, weak searching ability and easy to fall into local optimal solution. Therefore, a grey wolf optimization algorithm combining Levy flight and nonlinear inertia weights (LGWO) is proposed in this paper. The combination of Levy flight and nonlinear inertia weight is to improve the search efficiency and solve the problem that the search ability is weak and it is easy to fall into the local optimal solution. In summary, LGWO solves the problems of insufficient precision, poor robustness, weak searching ability and easy to fall into local optimal. This paper uses Congress on Evolutionary Computation benchmark function and combines algorithms with neural network for power line fault classification prediction to verify the effectiveness of each strategy improvement in LGWO and its comparison with other excellent algorithms (sine cosine algorithm, tree seed algorithm, wind driven optimization, and gravitational search algorithm). In the combination of neural networks and optimization algorithms, the accuracy of LGWO has been improved compared to the basic GWO, and LGWO has achieved the best performance in multiple algorithm comparisons. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. Deep learning techniques for Alzheimer's disease detection in 3D imaging: A systematic review.
- Author
-
Zia‐ur‐Rehman, Awang, Mohd Khalid, Ali, Ghulam, and Faheem, Muhammad
- Subjects
SUPERVISED learning ,ALZHEIMER'S disease ,RECURRENT neural networks ,CONVOLUTIONAL neural networks ,DEEP learning - Abstract
Background and Aims: Alzheimer's disease (AD) is a degenerative neurological condition that worsens over time and leads to deterioration in cognitive abilities, reduced memory, and, eventually, a decrease in overall functioning. Timely and correct identification of Alzheimer's is essential for effective treatment. The systematic study specifically examines the application of deep learning (DL) algorithms in identifying AD using three‐dimensional (3D) imaging methods. The main goal is to evaluate these methods' current state, efficiency, and potential enhancements, offering valuable insights into how DL could improve AD's rapid and precise diagnosis. Methods: We searched different online repositories, such as IEEE Xplore, Elsevier, MDPI, PubMed Central, Science Direct, ACM, Springer, and others, to thoroughly summarize current research on DL methods to diagnose AD by analyzing 3D imaging data published between 2020 and 2024. We use PRISMA (Preferred Reporting Items for Systematic Reviews and Meta‐Analyses) guidelines to ensure the organization and understandability of the information collection process. We thoroughly analyzed the literature to determine the primary techniques used in these investigations and their findings. Results and Conclusion: The ability of convolutional neural networks (CNNs) and their variations, including 3D CNNs and recurrent neural networks, to detect both temporal and spatial characteristics in volumetric data has led to their widespread use. Methods such as transfer learning, combining multimodal data, and using attention procedures have improved models' precision and reliability. We selected 87 articles for evaluation. Out of these, 31 papers included various concepts, explanations, and elucidations of models and theories, while the other 56 papers primarily concentrated on issues related to practical implementation. This article introduces popular imaging types, 3D imaging for Alzheimer's detection, discusses the benefits and restrictions of the DL‐based approach to AD assessment, and gives a view toward future developments resulting from critical evaluation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
45. Machine learning based fault detection technique for hybrid multi level inverter topology.
- Author
-
Chappa, Anilkumar, Rao, K. Dhananjay, Dhananjaya, Mudadla, Dawn, Subhojit, Al Mansur, Ahmed, and Ustun, Taha Selim
- Subjects
ELECTRIC inverters ,MACHINE learning ,ARTIFICIAL neural networks ,FEATURE extraction - Abstract
Multilevel inverters (MLIs) have a significant contribution in many industrial sectors due to their improved power quality and lesser voltage stress, over the conventional three‐level inverters. However, the implementation of MLIs with an increased device count creates the scope of development in MLIs topologies. In this regard, a hybrid MLI topology is studied in this paper whose architecture is based on conventional two‐level inverters. This topology has lesser device count characteristics when compared to conventional and most of the recently presented configurations for nine‐level output voltage generation. The major issue of capacitor voltage balancing is resolved by employing an appropriate switching strategy. However, the semiconductor switches are the most vulnerable components and causes the open circuit faults frequently that creates issues in real time operation. Hence, it is important to detect the open circuit fault in switches in the least possible time. A new approach to open circuit fault detection technique based on the analysis of load voltage waveform is proposed in this paper. The wavelet transform technique has been implemented for feature extraction of load voltage. Later, the classification of the fault has been achieved by training an artificial neural network (ANN). The proposed work has been studied in MATLAB/simulation and the obtained results are verified experimentally. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. MEMPSEP‐III. A Machine Learning‐Oriented Multivariate Data Set for Forecasting the Occurrence and Properties of Solar Energetic Particle Events Using a Multivariate Ensemble Approach.
- Author
-
Moreland, Kimberly, Dayeh, Maher A., Bain, Hazel M., Chatterjee, Subhamoy, Muñoz‐Jaramillo, Andrés, and Hart, Samuel T.
- Subjects
SOLAR energetic particles ,INTERPLANETARY magnetic fields ,SOLAR radio emission ,SPACE environment ,MACHINE learning ,CORONAL mass ejections ,SOLAR wind - Abstract
We introduce a new multivariate data set that utilizes multiple spacecraft collecting in‐situ and remote sensing heliospheric measurements shown to be linked to physical processes responsible for generating solar energetic particles (SEPs). Using the Geostationary Operational Environmental Satellites (GOES) flare event list from Solar Cycle (SC) 23 and part of SC 24 (1998–2013), we identify 252 solar events (>C‐class flares) that produce SEPs and 17,542 events that do not. For each identified event, we acquire the local plasma properties at 1 au, such as energetic proton and electron data, upstream solar wind conditions, and the interplanetary magnetic field vector quantities using various instruments onboard GOES and the Advanced Composition Explorer spacecraft. We also collect remote sensing data from instruments onboard the Solar Dynamic Observatory, Solar and Heliospheric Observatory, and the Wind solar radio instrument WAVES. The data set is designed to allow for variations of the inputs and feature sets for machine learning (ML) in heliophysics and has a specific purpose for forecasting the occurrence of SEP events and their subsequent properties. This paper describes a data set created from multiple publicly available observation sources that is validated, cleaned, and carefully curated for our ML pipeline. The data set has been used to drive the newly‐developed Multivariate Ensemble of Models for Probabilistic Forecast of SEPs (MEMPSEP; see MEMPSEP‐I (Chatterjee et al., 2024, https://doi.org/10.1029/2023SW003568) and MEMPSEP‐II (Dayeh et al., 2024, https://doi.org/10.1029/2023SW003697) for accompanying papers). Plain Language Summary: We present a new data set that uses observations from multiple spacecraft observing the Sun and the interplanetary space around it. This data is connected to the processes that create solar energetic particles (SEPs). SEP events pose threats to both astronauts and assets in space. The data set contains 252 solar flare events that caused SEPs and 17,542 that do not. For each event, we gather information about the local space environment around the sun, such as energetic protons and electrons, the conditions of the solar wind, the magnetic field, and remote solar imaging data. We use instruments from NOAA's Geostationary Operational Environmental Satellites (GOES) and the Advanced Composition Explorer spacecraft, as well as data from the Solar Dynamic Observatory, the Solar and Heliospheric Observatory, and the Wind solar radio instrument WAVES. This data set is designed to be used in machine learning (ML), with a focus on predicting the occurrence and properties of SEP events. We detail each observation obtained from publicly available sources, and the data treatment processes used to validate the reliability and usefulness for ML applications. Key Points: Machine learning oriented data set for predicting the occurance and properties of solar energetic particle eventsMultivariate remote sensing and in‐situ observationsContinuous data set spanning several decades [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
47. 32‐4: Late‐News Paper: OLED display Gamma LUT optimization based on Principal Component Analysis.
- Author
-
Kim, Hyunchul, Chang, Sun‐il, Youn, Sang Young, Foo, Ken, and Kim, Taesung
- Subjects
PRINCIPAL components analysis - Abstract
The high‐end OLED display has multiple gamma Look‐Up Tables(LUT) to minimize any noticeable luminance changes occuring in driving mode transition. We analyze redundancy of LUTs and propose visually lossless LUT compression based on Principal Component Analysis (PCA). It reduces costly One‐Time‐Programmable(OTP) memory size, typically used for LUT storage in Driver IC, by more than 50% with less than 0.1 Mean Square Error(MSE) for 10 bit LUTs. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
48. Making complex prediction rules applicable for readers: Current practice in random forest literature and recommendations.
- Author
-
Boulesteix, Anne‐Laure, Janitza, Silke, Hornung, Roman, Probst, Philipp, Busen, Hannah, and Hapfelmeier, Alexander
- Abstract
Ideally, prediction rules should be published in such a way that readers may apply them, for example, to make predictions for their own data. While this is straightforward for simple prediction rules, such as those based on the logistic regression model, this is much more difficult for complex prediction rules derived by machine learning tools. We conducted a survey of articles reporting prediction rules that were constructed using the random forest algorithm and published in PLOS ONE in 2014–2015 in the field "medical and health sciences", with the aim of identifying issues related to their applicability. Making a prediction rule reproducible is a possible way to ensure that it is applicable; thus reproducibility is also examined in our survey. The presented prediction rules were applicable in only 2 of 30 identified papers, while for further eight prediction rules it was possible to obtain the necessary information by contacting the authors. Various problems, such as nonresponse of the authors, hampered the applicability of prediction rules in the other cases. Based on our experiences from this illustrative survey, we formulate a set of recommendations for authors who aim to make complex prediction rules applicable for readers. All data including the description of the considered studies and analysis codes are available as supplementary materials. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
49. Multilayer self‐attention residual network for code search.
- Author
-
Hu, Haize, Liu, Jianxun, and Zhang, Xiangping
- Subjects
LINEAR network coding ,SOURCE code ,COMPUTER software development ,ACCURACY of information ,OPEN source software - Abstract
Summary: Software developers usually search existing code snippets in open source code repositories to modify and reuse them. Therefore, how to get the right code snippet from the open‐source code repository quickly and accurately is the focus of current software development research. Nowadays, code search is one of the solutions. To improve the accuracy of source code feature information representation and the accuracy of code search. A multilayer self‐ attention residual network‐based code search model (MSARN‐CS) is proposed in this paper. In the MSARN‐CS model, not only the weight of each word in the code sequence unit is considered but also the effect of embedding between code sequence units is calculated. In addition, an optimization model of residuals is introduced to compensate for the loss of information in the code sequences during the model training. To verify the search effectiveness of the MSARN‐CS model, three other baseline models are compared on the basis of extensive source code data. The experimental results show that the MSARN‐CS model has better search results compared with the baseline model. For parameter Recall@1, the experimental result of MSARN‐CS model was 9.547, which as 100.90%, 73.87%, 60.37%, and 2.55% better compared to CODEnn, CRLCS, SAN‐CS‐ and SAN‐CS, respectively. For the parameter Recall@5, the results improved by 26.67%, 36.23%, 36.21%, and 1.63%, respectively, and for the parameter Recall@10, the results improved by 13.92%, 25.70%, 20.78%, and 2.23%, respectively. For the parameter mean reciprocal rank, the results improved by 52.89%, 76.17%, 63.38%, and 3.88%, respectively. For the parameter normalized discounted cumulative gain, the results improved by 54.22%, 60.55%, 50.28%, and 3.30%, respectively. The MSARN‐CS model proposed in the paper can effectively improve the accuracy of code search and enhance the programming efficiency of developers. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
50. WHERE DO WE STAND IN CRYPTOCURRENCIES ECONOMIC RESEARCH? A SURVEY BASED ON HYBRID ANALYSIS.
- Author
-
Bariviera, Aurelio F. and Merediz‐Solà, Ignasi
- Subjects
ECONOMIC research ,CRYPTOCURRENCIES ,LITERATURE reviews ,MACHINE learning ,BIBLIOMETRICS - Abstract
This survey develops a dual analysis, consisting, first, in a bibliometric examination and, second, in a close literature review of all the scientific production around cryptocurrencies conducted in economics so far. The aim of this paper is twofold. On the one hand, proposes a methodological hybrid approach to perform comprehensive literature reviews. On the other hand, we provide an updated state of the art in cryptocurrency economic literature. Our methodology emerges as relevant when the topic comprises a large number of papers, which make unrealistic to perform a detailed reading of all the papers. This dual perspective offers a full landscape of cryptocurrency economic research. First, by means of the distant reading provided by machine learning bibliometric techniques, we are able to identify main topics, journals, key authors, and other macro aggregates. Second, based on the information provided by the previous stage, the traditional literature review provides a closer look at methodologies, data sources, and other details of the papers. In this way, we offer a classification and analysis of the mounting research produced in a relative short time span. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.