36 results on '"local interpretable model-agnostic explanations"'
Search Results
2. Estimating stay cable vibration under typhoon with an explainable ensemble learning model.
- Author
-
Sun, Zhen, Ye, Xiao-Wei, and Lu, Jun
- Subjects
- *
MACHINE learning , *STRUCTURAL health monitoring , *VIBRATION (Mechanics) , *ACCELERATION (Mechanics) , *REGRESSION trees , *TYPHOONS , *CABLE-stayed bridges - Abstract
Excessive vibration of stay cables in strong winds has been a concern for bridge operators, which impairs the durability of both the cables and the bridge structure. This paper develops a data-driven approach to predict the amplitude of the cable vibration using an ensemble learning model. The model aims to predict cable vibrations in both in-plane and out-of-plane directions, with the wind speed, wind direction, turbulence intensity, and deck acceleration as input variables. Especially, the deck acceleration is included considering the deck-cable interaction and vehicle effects, which significantly improved the accuracy of the prediction. Furthermore, the model is interpreted with local interpretable model-agnostic explanations (LIME) and partial dependence plot (PDP) methods. The former demonstrates the relative importance of input variables on a global scale, and the latter indicates the correlation between individual input variables with the prediction target. The investigation is validated using the data harnessed from structural health monitoring (SHM) of a 1088-m cable-stayed bridge during three typhoon events. The adopted Gradient boosting regression tree (GBRT) model demonstrated better performance than other state-of-the-art machine learning models. The developed approach can provide guidance on preventive maintenance of stay cables to avoid damage due to excessive vibration. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. An Explainable Deep-Learning Model to Aid in the Diagnosis of Age Related Macular Degeneration
- Author
-
Herrero-Tudela, María, Romero-Oraá, Roberto, Hornero, Roberto, Gutiérrez-Tobal, Gonzalo C., Lopez, María I., García, María, Magjarević, Ratko, Series Editor, Ładyżyński, Piotr, Associate Editor, Ibrahim, Fatimah, Associate Editor, Lackovic, Igor, Associate Editor, Rock, Emilio Sacristan, Associate Editor, Jarm, Tomaž, editor, Šmerc, Rok, editor, and Mahnič-Kalamiza, Samo, editor
- Published
- 2024
- Full Text
- View/download PDF
4. Improving Deep Learning Transparency: Leveraging the Power of LIME Heatmap
- Author
-
Farhood, Helia, Najafi, Mohammad, Saberi, Morteza, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Monti, Flavia, editor, Plebani, Pierluigi, editor, Moha, Naouel, editor, Paik, Hye-young, editor, Barzen, Johanna, editor, Ramachandran, Gowri, editor, Bianchini, Devis, editor, Tamburri, Damian A., editor, and Mecella, Massimo, editor
- Published
- 2024
- Full Text
- View/download PDF
5. Foreign direct investment and local interpretable model-agnostic explanations: a rational framework for FDI decision making
- Author
-
Singh, Devesh
- Published
- 2024
- Full Text
- View/download PDF
6. Foreign direct investment and local interpretable model-agnostic explanations: a rational framework for FDI decision making
- Author
-
Devesh Singh
- Subjects
FDI ,Machine learning ,Interpretable machine learning ,Local interpretable model-agnostic explanations ,Business ,HF5001-6182 - Abstract
Purpose – This study aims to examine foreign direct investment (FDI) factors and develops a rational framework for FDI inflow in Western European countries such as France, Germany, the Netherlands, Switzerland, Belgium and Austria. Design/methodology/approach – Data for this study were collected from the World development indicators (WDI) database from 1995 to 2018. Factors such as economic growth, pollution, trade, domestic capital investment, gross value-added and the financial stability of the country that influence FDI decisions were selected through empirical literature. A framework was developed using interpretable machine learning (IML), decision trees and three-stage least squares simultaneous equation methods for FDI inflow in Western Europe. Findings – The findings of this study show that there is a difference between the most important and trusted factors for FDI inflow. Additionally, this study shows that machine learning (ML) models can perform better than conventional linear regression models. Research limitations/implications – This research has several limitations. Ideally, classification accuracies should be higher, and the current scope of this research is limited to examining the performance of FDI determinants within Western Europe. Practical implications – Through this framework, the national government can understand how investors make their capital allocation decisions in their country. The framework developed in this study can help policymakers better understand the rationality of FDI inflows. Originality/value – An IML framework has not been developed in prior studies to analyze FDI inflows. Additionally, the author demonstrates the applicability of the IML framework for estimating FDI inflows in Western Europe.
- Published
- 2024
- Full Text
- View/download PDF
7. Interpretable machine learning model for early prediction of 28-day mortality in ICU patients with sepsis-induced coagulopathy: development and validation
- Author
-
Shu Zhou, Zongqing Lu, Yu Liu, Minjie Wang, Wuming Zhou, Xuanxuan Cui, Jin Zhang, Wenyan Xiao, Tianfeng Hua, Huaqing Zhu, and Min Yang
- Subjects
Sepsis induced coagulopathy ,Gradient boosting decision tree ,Machine learning ,Shapley additive explanations ,Local interpretable model-agnostic explanations ,Medicine - Abstract
Abstract Objective Sepsis-induced coagulopathy (SIC) is extremely common in individuals with sepsis, significantly associated with poor outcomes. This study attempted to develop an interpretable and generalizable machine learning (ML) model for early predicting the risk of 28-day death in patients with SIC. Methods In this retrospective cohort study, we extracted SIC patients from the Medical Information Mart for Intensive Care III (MIMIC-III), MIMIC-IV, and eICU-CRD database according to Toshiaki Iba's scale. And the overlapping in the MIMIC-IV was excluded for this study. Afterward, only the MIMIC-III cohort was randomly divided into the training set, and the internal validation set according to the ratio of 7:3, while the MIMIC-IV and eICU-CRD databases were considered the external validation sets. The predictive factors for 28-day mortality of SIC patients were determined using recursive feature elimination combined with tenfold cross-validation (RFECV). Then, we constructed models using ML algorithms. Multiple metrics were used for evaluation of performance of the models, including the area under the receiver operating characteristic curve (AUROC), area under the precision recall curve (AUPRC), accuracy, sensitivity, specificity, negative predictive value, positive predictive value, recall, and F1 score. Finally, Shapley Additive Explanations (SHAP), Local Interpretable Model-Agnostic Explanations (LIME) were employed to provide a reasonable interpretation for the prediction results. Results A total of 3280, 2798, and 1668 SIC patients were screened from MIMIC-III, MIMIC-IV, and eICU-CRD databases, respectively. Seventeen features were selected to construct ML prediction models. XGBoost had the best performance in predicting the 28-day mortality of SIC patients, with AUC of 0.828, 0.913 and 0.923, the AUPRC of 0.807, 0.796 and 0.921, the accuracy of 0.785, 0.885 and 0.891, the F1 scores were 0.63, 0.69 and 0.70 in MIMIC-III (internal validation set), MIMIC-IV, and eICU-CRD databases. The importance ranking and SHAP analyses showed that initial SOFA score, red blood cell distribution width (RDW), and age were the top three critical features in the XGBoost model. Conclusions We developed an optimal and explainable ML model to predict the risk of 28-day death of SIC patients 28-day death risk. Compared with conventional scoring systems, the XGBoost model performed better. The model established will have the potential to improve the level of clinical practice for SIC patients. Graphical Abstract
- Published
- 2024
- Full Text
- View/download PDF
8. Explainable AI for Intrusion Detection Systems: LIME and SHAP Applicability on Multi-Layer Perceptron
- Author
-
Diogo Gaspar, Paulo Silva, and Catarina Silva
- Subjects
Artificial intelligence ,explainability ,intrusion detection system ,local interpretable model-agnostic explanations ,machine learning ,shapley additive explanations ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Machine learning-based systems have presented increasing learning performance, in a wide variety of tasks. However, the problem with some state-of-the-art models is their lack of transparency, trustworthiness, and explainability. To address this problem, eXplainable Artificial Intelligence (XAI) appeared. It is a research field that aims to make black-box models more understandable to humans. The research on this topic has increased in recent years, and many methods, such as LIME (Local Interpretable Model-Agnostic Explanations) and SHAP (SHapley Additive exPlanations) have been proposed. Machine learning-based Intrusion Detection Systems (IDS) are one of the many application domains of XAI. However, most of the works about model interpretation focus on other fields, like computer vision, natural language processing, biology, healthcare, etc. This poses a challenge for cybersecurity professionals tasked with analyzing IDS results, thereby impeding their capacity to make informed decisions. In an attempt to address this problem, we have selected two XAI methods, LIME, and SHAP. Using the methods, we have retrieved explanations for the results of a black-box model, part of an IDS solution that performs intrusion detection on IoT devices, increasing its interpretability. In order to validate the explanations, we carried out a perturbation analysis where we tried to obtain a different classification based on the features present in the explanations. With the explanations and the perturbation analysis we were able to draw conclusions about the negative impact of particular features on the model results when present in the input data, making it easier for cybersecurity experts when analyzing the model results and it serves as an aid to the continuous improvement the model. The perturbations also serve as a comparison of performance between LIME and SHAP. To evaluate the degree of interpretability increase, and the explanations provided by each XAI method of the model and directly compare the XAI methods, we have performed a survey analysis.
- Published
- 2024
- Full Text
- View/download PDF
9. Interpretable machine learning model for early prediction of 28-day mortality in ICU patients with sepsis-induced coagulopathy: development and validation.
- Author
-
Zhou, Shu, Lu, Zongqing, Liu, Yu, Wang, Minjie, Zhou, Wuming, Cui, Xuanxuan, Zhang, Jin, Xiao, Wenyan, Hua, Tianfeng, Zhu, Huaqing, and Yang, Min
- Subjects
MACHINE learning ,RECEIVER operating characteristic curves ,ERYTHROCYTES ,BLOOD coagulation disorders ,PREDICTION models - Abstract
Objective: Sepsis-induced coagulopathy (SIC) is extremely common in individuals with sepsis, significantly associated with poor outcomes. This study attempted to develop an interpretable and generalizable machine learning (ML) model for early predicting the risk of 28-day death in patients with SIC. Methods: In this retrospective cohort study, we extracted SIC patients from the Medical Information Mart for Intensive Care III (MIMIC-III), MIMIC-IV, and eICU-CRD database according to Toshiaki Iba's scale. And the overlapping in the MIMIC-IV was excluded for this study. Afterward, only the MIMIC-III cohort was randomly divided into the training set, and the internal validation set according to the ratio of 7:3, while the MIMIC-IV and eICU-CRD databases were considered the external validation sets. The predictive factors for 28-day mortality of SIC patients were determined using recursive feature elimination combined with tenfold cross-validation (RFECV). Then, we constructed models using ML algorithms. Multiple metrics were used for evaluation of performance of the models, including the area under the receiver operating characteristic curve (AUROC), area under the precision recall curve (AUPRC), accuracy, sensitivity, specificity, negative predictive value, positive predictive value, recall, and F1 score. Finally, Shapley Additive Explanations (SHAP), Local Interpretable Model-Agnostic Explanations (LIME) were employed to provide a reasonable interpretation for the prediction results. Results: A total of 3280, 2798, and 1668 SIC patients were screened from MIMIC-III, MIMIC-IV, and eICU-CRD databases, respectively. Seventeen features were selected to construct ML prediction models. XGBoost had the best performance in predicting the 28-day mortality of SIC patients, with AUC of 0.828, 0.913 and 0.923, the AUPRC of 0.807, 0.796 and 0.921, the accuracy of 0.785, 0.885 and 0.891, the F
1 scores were 0.63, 0.69 and 0.70 in MIMIC-III (internal validation set), MIMIC-IV, and eICU-CRD databases. The importance ranking and SHAP analyses showed that initial SOFA score, red blood cell distribution width (RDW), and age were the top three critical features in the XGBoost model. Conclusions: We developed an optimal and explainable ML model to predict the risk of 28-day death of SIC patients 28-day death risk. Compared with conventional scoring systems, the XGBoost model performed better. The model established will have the potential to improve the level of clinical practice for SIC patients. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
10. Explainable prediction of loan default based on machine learning models
- Author
-
Xu Zhu, Qingyong Chu, Xinchang Song, Ping Hu, and Lu Peng
- Subjects
Explainable prediction ,Machine learning ,Loan default ,Local interpretable model-agnostic explanations ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Owing to the convenience of online loans, an increasing number of people are borrowing money on online platforms. With the emergence of machine learning technology, predicting loan defaults has become a popular topic. However, machine learning models have a black-box problem that cannot be disregarded. To make the prediction model rules more understandable and thereby increase the user’s faith in the model, an explanatory model must be used. Logistic regression, decision tree, XGBoost, and LightGBM models are employed to predict a loan default. The prediction results show that LightGBM and XGBoost outperform logistic regression and decision tree models in terms of the predictive ability. The area under curve for LightGBM is 0.7213. The accuracies of LightGBM and XGBoost exceed 0.8. The precisions of LightGBM and XGBoost exceed 0.55. Simultaneously, we employed the local interpretable model-agnostic explanations approach to undertake an explainable analysis of the prediction findings. The results show that factors such as the loan term, loan grade, credit rating, and loan amount affect the predicted outcomes.
- Published
- 2023
- Full Text
- View/download PDF
11. Explainable Decision Tree-Based Screening of Cognitive Impairment Leveraging Minimal Neuropsychological Tests
- Author
-
Poonam, Km, Prasad, Aayush, Guha, Rajlakshmi, Hazra, Aritra, Chakrabarti, Partha P., Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Maji, Pradipta, editor, Huang, Tingwen, editor, Pal, Nikhil R., editor, Chaudhury, Santanu, editor, and De, Rajat K., editor
- Published
- 2023
- Full Text
- View/download PDF
12. Explainable AI for Intrusion Prevention: A Review of Techniques and Applications
- Author
-
Chandre, Pankaj R., Vanarote, Viresh, Patil, Rajkumar, Mahalle, Parikshit N., Shinde, Gitanjali R., Nimbalkar, Madhukar, Barot, Janki, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Choudrie, Jyoti, editor, Mahalle, Parikshit N., editor, Perumal, Thinagaran, editor, and Joshi, Amit, editor
- Published
- 2023
- Full Text
- View/download PDF
13. Frontotemporal Dementia Detection Model Based on Explainable Machine Learning Approach
- Author
-
Poonam, Km, Guha, Rajlakshmi, Chakrabarti, Partha P., Rannenberg, Kai, Editor-in-Chief, Soares Barbosa, Luís, Editorial Board Member, Goedicke, Michael, Editorial Board Member, Tatnall, Arthur, Editorial Board Member, Neuhold, Erich J., Editorial Board Member, Stiller, Burkhard, Editorial Board Member, Stettner, Lukasz, Editorial Board Member, Pries-Heje, Jan, Editorial Board Member, Kreps, David, Editorial Board Member, Rettberg, Achim, Editorial Board Member, Furnell, Steven, Editorial Board Member, Mercier-Laurent, Eunika, Editorial Board Member, Winckler, Marco, Editorial Board Member, Malaka, Rainer, Editorial Board Member, Chandran K R, Sarath, editor, N, Sujaudeen, editor, A, Beulah, editor, and Hamead H, Shahul, editor
- Published
- 2023
- Full Text
- View/download PDF
14. Contextualized Embeddings from Transformers for Sentiment Analysis on Code-Mixed Hinglish Data: An Expanded Approach with Explainable Artificial Intelligence
- Author
-
Yadav, Sargam, Kaushik, Abhishek, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, M, Anand Kumar, editor, Chakravarthi, Bharathi Raja, editor, B, Bharathi, editor, O’Riordan, Colm, editor, Murthy, Hema, editor, Durairaj, Thenmozhi, editor, and Mandl, Thomas, editor
- Published
- 2023
- Full Text
- View/download PDF
15. Explainable artificial intelligence for intrusion detection in IoT networks: A deep learning based approach
- Author
-
Sharma, Bhawana (author), Sharma, Lokesh (author), Lal, C. (author), Roy, Satyabrata (author), Sharma, Bhawana (author), Sharma, Lokesh (author), Lal, C. (author), and Roy, Satyabrata (author)
- Abstract
The Internet of Things (IoT) is currently seeing tremendous growth due to new technologies and big data. Research in the field of IoT security is an emerging topic. IoT networks are becoming more vulnerable to new assaults as a result of the growth in devices and the production of massive data. In order to recognize the attacks, an intrusion detection system is required. In this work, we suggested a Deep Learning (DL) model for intrusion detection to categorize various attacks in the dataset. We used a filter-based approach to pick out the most important aspects and limit the number of features, and we built two different deep-learning models for intrusion detection. For model training and testing, we used two publicly accessible datasets, NSL-KDD and UNSW-NB 15. First, we applied the dataset on the Deep neural network (DNN) model and then the same dataset on Convolution Neural Network (CNN) model. For both datasets, the DL model had a better accuracy rate. Because DL models are opaque and challenging to comprehend, we applied the idea of explainable Artificial Intelligence (AI) to provide a model explanation. To increase confidence in the DNN model, we applied the explainable AI (XAI) Local Interpretable Model-agnostic Explanations (LIME ) method, and for better understanding, we also applied Shapley Additive Explanations (SHAP)., Green Open Access added to TU Delft Institutional Repository ‘You share, we take care!’ – Taverne project https://www.openaccess.nl/en/you-share-we-take-care Otherwise as indicated in the copyright section: the publisher is the copyright holder of this work and the author uses the Dutch legislation to make this work public., Cyber Security
- Published
- 2024
- Full Text
- View/download PDF
16. An efficient interpretable stacking ensemble model for lung cancer prognosis.
- Author
-
Arif U, Zhang C, Hussain S, and Abbasi AR
- Abstract
Lung cancer significantly contributes to global cancer mortality, posing challenges in clinical management. Early detection and accurate prognosis are crucial for improving patient outcomes. This study develops an interpretable stacking ensemble model (SEM) for lung cancer prognosis prediction and identifies key risk factors. Using a Kaggle dataset of 1000 patients with 22 variables, the model classifies prognosis into Low, Medium, and High-risk categories. The bootstrap method was employed for evaluation metrics, while SHAP (Shapley Additive Explanations) and LIME (Local Interpretable Model-agnostic Explanations) assessed model interpretability. Results showed SEM's superior interpretability over traditional models, such as Random Forest, Logistic Regression, Decision Tree, Gradient Boosting Machine, Extreme Gradient Boosting Machine, and Light Gradient Boosting Machine. SEM achieved an accuracy of 98.90 %, precision of 98.70 %, F1 score of 98.85 %, sensitivity of 98.77 %, specificity of 95.45 %, Cohen's kappa value of 94.56 %, and an AUC of 98.10 %. The SEM demonstrated robust performance in lung cancer prognosis, revealing chronic lung cancer and genetic risk as major factors., Competing Interests: Declaration of Competing Interest All authors declared that they have no conflict of interest., (Copyright © 2024 Elsevier Ltd. All rights reserved.)
- Published
- 2024
- Full Text
- View/download PDF
17. Interpretable Machine-Learning Approach in Estimating FDI Inflow: Visualization of ML Models with LIME and H2O
- Author
-
Singh Devesh
- Subjects
fdi ,h2o ,local interpretable model-agnostic explanations ,machine learning ,Political science ,International relations ,JZ2-6530 - Abstract
In advancement of interpretable machine learning (IML), this research proposes local interpretable model-agnostic explanations (LIME) as a new visualization technique in a novel informative way to analyze the foreign direct investment (FDI) inflow. This article examines the determinants of FDI inflow through IML with a supervised learning method to analyze the foreign investment determinants in Hungary by using an open-source artificial intelligence H2O platform. This author used three ML algorithms—general linear model (GML), gradient boosting machine (GBM), and random forest (RF) classifier—to analyze the FDI inflow from 2001 to 2018. The result of this study shows that in all three classifiers GBM performs better to analyze FDI inflow determinants. The variable value of production in a region is the most influenced determinant to the inflow of FDI in Hungarian regions. Explanatory visualizations are presented from the analyzed dataset, which leads to their use in decision-making.
- Published
- 2021
- Full Text
- View/download PDF
18. Explainable Machine Learning Approach Quantified the Long-Term (1981–2015) Impact of Climate and Soil Properties on Yields of Major Agricultural Crops Across CONUS
- Author
-
Debjani Sihi, Biswanath Dari, Abraham Peedikayil Kuruvila, Gaurav Jha, and Kanad Basu
- Subjects
climate change ,crop production ,environment ,soils ,explainable machine learning ,local interpretable model-agnostic explanations ,Nutrition. Foods and food supply ,TX341-641 ,Food processing and manufacture ,TP368-456 - Abstract
A comprehensive understanding of the long-term data on the crop, soils, environment, climate, and production management would facilitate efficient data-driven decision-making in agriculture production under changing climate. We have employed an explainable machine learning algorithm (random forest model coupled with LIME; Local Interpretable Model-Agnostic Explanations framework) using multi-decadal (1981–2015) data on climate variables, soil properties, and yield of major crops across the Coterminous United States (CONUS). This data-driven approach explained the multi-faceted factors of crop production for corn, soybean, cotton, and wheat under field conditions by leveraging agricultural informatics. We attempted to show how crop yields can better be correlated and explained when production input varies along with changing climatic/environmental and edaphic conditions. Our findings suggest Growing Degree Days (GDDs) as important climatic factors, while water holding capacity is one of the dominant soil properties in interpreting crop yield variability. Our findings will facilitate growers, crop production scientists, land management specialists, stakeholders, and policy makers in their future decision-making processes related to sustainable and long-term soil, water, and crop management practices.
- Published
- 2022
- Full Text
- View/download PDF
19. Stratifying the Risk of Cardiovascular Disease in Obstructive Sleep Apnea Using Machine Learning.
- Author
-
Gourishetti, Saikrishna C., Taylor, Rodney, and Isaiah, Amal
- Abstract
Objectives/Hypothesis: Obstructive sleep apnea (OSA) is associated with higher risk of morbidity and mortality related to cardiovascular disease (CVD). Due to overlapping clinical risk factors, identifying high‐risk patients with OSA who are likely to develop CVD remains challenging. We aimed to identify baseline clinical factors associated with the future development of CVD in patients with OSA. Study Design: Retrospective analysis of prospectively collected data. Methods: We performed a retrospective analysis of 967 adults aged 45 to 84 years and enrolled in the Multi‐Ethnic Study of Atherosclerosis. Six machine learning models were created using baseline clinical factors initially identified by stepwise variable selection. The performance of these models for the prediction of additional risk of CVD in OSA was calculated. Additionally, these models were evaluated for interpretability using locally interpretable model‐agnostic explanations. Results: Of the 967 adults without baseline OSA or CVD, 116 were diagnosed with OSA and CVD and 851 with OSA alone 10 years after enrollment. The best performing models included random forest (sensitivity 84%, specificity 99%, balanced accuracy 91%) and bootstrap aggregation (sensitivity 84%, specificity 100%, balanced accuracy 92%). The strongest predictors of OSA and CVD versus OSA alone were fasting glucose >91 mg/dL, diastolic pressure >73 mm Hg, and age >59 years. Conclusion: In the selected study population of adults without OSA or CVD at baseline, the strongest predictors of CVD in patients with OSA include fasting glucose, diastolic pressure, and age. These results may shape a strategy for cardiovascular risk stratification in patients with OSA and early intervention to mitigate CVD‐related morbidity. Level of Evidence: 3 Laryngoscope, 132:234–241, 2022 [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
20. Explainable Soft Attentive EfficientNet for breast cancer classification in histopathological images.
- Author
-
Peta, Jyothi and Koppu, Srinivas
- Subjects
TUMOR classification ,IMAGE recognition (Computer vision) ,BREAST ,BREAST cancer ,TIME complexity ,FALSE discovery rate - Abstract
Breast Cancer (BC) is believed to be the cancer that occurs most frequently in women worldwide, taking the lives of it's the victims. In early diagnosis aids the patients to survive under greater probability. Several existing studies utilize diagnostic mechanisms via histopathology image for early identification of breast tumors. However, it increases the medical costs and consumes the time. Thus, in order to accurately classify the breast tumor, this study suggests a novel explainable DL technique. Using this technique, better accuracy is achieved while performing classifications. Improved accuracy may greatly help the medical practitioners for classifying breast cancer effectively. Initially, adaptive unsharp mask filtering (AUMF) technique is proposed to remove the noise and enhance the quality of the image. Finally, Explainable Soft Attentive EfficientNet (ESAE-Net) technique is introduced to classify the breast tumor (BT). Four explainable algorithms are investigated for improved visualizations over the BTs: Gradient-Weighted Class Activation Mapping (Grad-CAM) Shapley additive explanations (SHAP), Contextual Importance and Utility (CIU), and Local Interpretable Model-Agnostic Explanations (LIME). The suggested approach uses two publicly accessible images of breast histopathology and is carried out on a Python platform. Performance metrics such as time complexity, False Discovery Rate (FDR), accuracy, and Mathew's correlation coefficient (MCC) are examined and contrasted with traditional research. In the experimental section, the proposed obtains an accuracy of 97.85% for dataset 1 and accuracy of 98.05% for dataset 2. In comparison with other existing methods, the proposed method is more efficient while using ESAE-Net for classifying the Breast cancer. • A novel and effective XAI based breast cancer (BC) classification framework (ESAE-Net) that can provide better learning interpretations and decision making processes effectively. • To introduce a novel adaptive unsharp mask filtering (AUMF) technique for enhancing the quality of the breast histopathology image. • To propose a new and efficient Explainable Soft Attentive EfficientNet (ESAE-Net) model to deal with the classification of BC from histopathology images. • To improve the proposed ESAE-Net by introducing various interpretability procedures like LIME, SHAP, CIU and Grad-CAM. • To validate the proposed method by carrying out extensive simulations under different simulation scenarios to prove the effectiveness of the proposed framework compared to other XAI frameworks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. An Explainable Artificial Intelligence Framework for the Deterioration Risk Prediction of Hepatitis Patients.
- Author
-
Peng, Junfeng, Zou, Kaiqiang, Zhou, Mi, Teng, Yi, Zhu, Xiongyong, Zhang, Feifei, and Xu, Jun
- Subjects
- *
HEPATITIS diagnosis , *DEEP learning , *DECISION trees , *SUPPORT vector machines , *ARTIFICIAL intelligence , *HEPATITIS , *RANDOM forest algorithms , *RISK assessment , *DESCRIPTIVE statistics , *RESEARCH funding , *COMPUTER-aided diagnosis , *DECISION making in clinical medicine , *LOGISTIC regression analysis , *STATISTICAL models - Abstract
In recent years, artificial intelligence-based computer aided diagnosis (CAD) system for the hepatitis has made great progress. Especially, the complex models such as deep learning achieve better performance than the simple ones due to the nonlinear hypotheses of the real world clinical data. However,complex model as a black box, which ignores why it make a certain decision, causes the model distrust from clinicians. To solve these issues, an explainable artificial intelligence (XAI) framework is proposed in this paper to give the global and local interpretation of auxiliary diagnosis of hepatitis while retaining the good prediction performance. First, a public hepatitis classification benchmark from UCI is used to test the feasibility of the framework. Then, the transparent and black-box machine learning models are both employed to forecast the hepatitis deterioration. The transparent models such as logistic regression (LR), decision tree (DT)and k-nearest neighbor (KNN) are picked. While the black-box model such as the eXtreme Gradient Boosting (XGBoost), support vector machine (SVM), random forests (RF) are selected. Finally, the SHapley Additive exPlanations (SHAP), Local Interpretable Model-agnostic Explanations (LIME) and Partial Dependence Plots (PDP) are utilized to improve the model interpretation of liver disease. The experimental results show that the complex models outperform the simple ones. The developed RF achieves the highest accuracy (91.9%) among all the models. The proposed framework combining the global and local interpretable methods improves the transparency of complex models, and gets insight into the judgments from the complex models, thereby guiding the treatment strategy and improving the prognosis of hepatitis patients. In addition, the proposed framework could also assist the clinical data scientists to design a more appropriate structure of CAD. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
22. Toward better prediction of recurrence for Cushing's disease: a factorization-machine based neural approach.
- Author
-
Fan, Yanghua, Li, Dongfang, Liu, Yifan, Feng, Ming, Chen, Qingcai, and Wang, Renzhi
- Abstract
Cushing's disease (CD) is a rare disease that occurs in 1.2–1.4 persons per million population per year. Recurrence prediction after transsphenoidal surgery (TSS) is important for determining individual treatment and follow-up strategies. Between 2000 and 2017, 354 CD patients with initial postoperative remission and long-term follow-up data were enrolled from Peking union medical college hospital (PUMCH) to predict recurrence, and PUMCH is one of the largest CD treatment centers in the world. We first investigated the effect of a factorization machine (FM)-based neural network to predict recurrence after TSS for CD. This method could automatically reduce a portion of the cross-feature selection work with acceptable parameters. We conducted a performance comparison of various algorithms on the collected dataset. To address the lack of interpretability of neural network models, we also used the local interpretable model-agnostic explanations approach, which provides an explanation in the form of relevant features of the predicted results by approximating the model behavior of the variables in a local manner. Compared with existing methods, the DeepFM model obtained the highest AUC value (0.869) and the lowest log loss value (0.256). According to the importance of each feature, three top features for the DeepFM model were postoperative morning adrenocorticotropic hormone level, age, and postoperative morning serum cortisol nadir. In the post hoc explanation phase, the above-mentioned importance-leading features made a great contribution to the prediction probability. The results showed that deep learning-based models could better aid neurosurgeons in recurrence prediction after TTS for patients with CD, and could contribute to determining individual treatment strategies. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
23. Justifying Short-Term Load Forecasts Obtained with the Use of Neural Models
- Author
-
Tadeusz A. Grzeszczyk and Michal K. Grzeszczyk
- Subjects
time-series forecasting ,short-term load forecasting ,energy forecasting model ,neural networks ,explainability ,local interpretable model-agnostic explanations ,Technology - Abstract
There is a lot of research on the neural models used for short-term load forecasting (STLF), which is crucial for improving the sustainable operation of energy systems with increasing technical, economic, and environmental requirements. Neural networks are computationally powerful; however, the lack of clear, readable and trustworthy justification of STLF obtained using such models is a serious problem that needs to be tackled. The article proposes an approach based on the local interpretable model-agnostic explanations (LIME) method that supports reliable premises justifying and explaining the forecasts. The use of the proposed approach makes it possible to improve the reliability of heuristic and experimental neural modeling processes, the results of which are difficult to interpret. Explaining the forecasting may facilitate the justification of the selection and the improvement of neural models for STLF, while contributing to a better understanding of the obtained results and broadening the knowledge and experience supporting the enhancement of energy systems security based on reliable forecasts and simplifying dispatch decisions.
- Published
- 2022
- Full Text
- View/download PDF
24. Interpretable Machine-Learning Approach in Estimating FDI Infl ow: Visualization of ML Models with LIME and H2O.
- Author
-
Singh, Devesh
- Subjects
SUPERVISED learning ,FOREIGN investments ,VISUALIZATION ,MACHINE learning ,RANDOM forest algorithms - Abstract
In advancement of interpretable machine learning (IML), this research proposes local interpretable model-agnostic explanations (LIME) as a new visualization technique in a novel informative way to analyze the foreign direct investment (FDI) infl ow. This article examines the determinants of FDI infl ow through IML with a supervised learning method to analyze the foreign investment determinants in Hungary by using an open-source artifi cial intelligence H2O platform. This author used three ML algorithms--general linear model (GML), gradient boosting machine (GBM), and random forest (RF) classifi er--to analyze the FDI infl ow from 2001 to 2018. The result of this study shows that in all three classifi ers GBM performs better to analyze FDI infl ow determinants. The variable value of production in a region is the most infl uenced determinant to the infl ow of FDI in Hungarian regions. Explanatory visualizations are presented from the analyzed dataset, which leads to their use in decision-making. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
25. Comprehensible Machine-Learning-Based Models for the Pre-Emptive Diagnosis of Multiple Sclerosis Using Clinical Data: A Retrospective Study in the Eastern Province of Saudi Arabia
- Author
-
Sunday O. Olatunji, Nawal Alsheikh, Lujain Alnajrani, Alhatoon Alanazy, Meshael Almusairii, Salam Alshammasi, Aisha Alansari, Rim Zaghdoud, Alaa Alahmadi, Mohammed Imran Basheer Ahmed, Mohammed Salih Ahmed, and Jamal Alhiyafi
- Subjects
shapley additive explanation ,local interpretable model-agnostic explanations ,machine learning ,explainable artificial intelligence ,Health, Toxicology and Mutagenesis ,pre-emptive diagnosis ,Public Health, Environmental and Occupational Health ,multiple sclerosis - Abstract
Multiple Sclerosis (MS) is characterized by chronic deterioration of the nervous system, mainly the brain and the spinal cord. An individual with MS develops the condition when the immune system begins attacking nerve fibers and the myelin sheathing that covers them, affecting the communication between the brain and the rest of the body and eventually causing permanent damage to the nerve. Patients with MS (pwMS) might experience different symptoms depending on which nerve was damaged and how much damage it has sustained. Currently, there is no cure for MS; however, there are clinical guidelines that help control the disease and its accompanying symptoms. Additionally, no specific laboratory biomarker can precisely identify the presence of MS, leaving specialists with a differential diagnosis that relies on ruling out other possible diseases with similar symptoms. Since the emergence of Machine Learning (ML) in the healthcare industry, it has become an effective tool for uncovering hidden patterns that aid in diagnosing several ailments. Several studies have been conducted to diagnose MS using ML and Deep Learning (DL) models trained using MRI images, achieving promising results. However, complex and expensive diagnostic tools are needed to collect and examine imaging data. Thus, the intention of this study is to implement a cost-effective, clinical data-driven model that is capable of diagnosing pwMS. The dataset was obtained from King Fahad Specialty Hospital (KFSH) in Dammam, Saudi Arabia. Several ML algorithms were compared, namely Support Vector Machine (SVM), Decision Tree (DT), Logistic Regression (LR), Random Forest (RF), Extreme Gradient Boosting (XGBoost), Adaptive Boosting (AdaBoost), and Extra Trees (ET). The results indicated that the ET model outpaced the rest with an accuracy of 94.74%, recall of 97.26%, and precision of 94.67%.
- Published
- 2023
- Full Text
- View/download PDF
26. Explainable Reinforcement Learning for Gameplay
- Author
-
Costa Sánchez, Àlex and Costa Sánchez, Àlex
- Abstract
State-of-the-art Machine Learning (ML) algorithms show impressive results for a myriad of applications. However, they operate as a sort of a black box: the decisions taken are not human-understandable. There is a need for transparency and interpretability of ML predictions to be wider accepted in society, especially in specific fields such as medicine or finance. Most of the efforts so far have focused on explaining supervised learning. This project aims to use some of these successful explainability algorithms and apply them to Reinforcement Learning (RL). To do so, we explain the actions of a RL agent playing Atari’s Breakout game, using two different explainability algorithms: Shapley Additive Explanations (SHAP) and Local Interpretable Model-agnostic Explanations (LIME). We successfully implement both algorithms, which yield credible and insightful explanations of the mechanics of the agent. However, we think the final presentation of the results is sub-optimal for the final user, as it is not intuitive at first sight., De senaste algoritmerna för maskininlärning (ML) visar imponerande resultat för en mängd olika tillämpningar. De fungerar dock som ett slags ”svart låda”: de beslut som fattas är inte begripliga för människor. Det finns ett behov av öppenhet och tolkningsbarhet för ML-prognoser för att de ska bli mer accepterade i samhället, särskilt inom specifika områden som medicin och ekonomi. De flesta insatser hittills har fokuserat på att förklara övervakad inlärning. Syftet med detta projekt är att använda några av dessa framgångsrika algoritmer för att förklara och tillämpa dem på förstärkning lärande (Reinforcement Learning, RL). För att göra detta förklarar vi handlingarna hos en RL-agent som spelar Ataris Breakout-spel med hjälp av två olika förklaringsalgoritmer: Shapley Additive Explanations (SHAP) och Local Interpretable Model-agnostic Explanations (LIME). Vi genomför framgångsrikt båda algoritmerna, som ger trovärdiga och insiktsfulla förklaringar av agentens mekanik. Vi anser dock att den slutliga presentationen av resultaten inte är optimal för slutanvändaren, eftersom den inte är intuitiv vid första anblicken., Els algoritmes d’aprenentatge automàtic (Machine Learning, ML) d’última generació mostren resultats impressionants per a moltes aplicacions. Tot i això, funcionen com una mena de caixa negra: les decisions preses no són comprensibles per a l’ésser humà. Per tal que les prediccion preses mitjançant ML siguin més acceptades a la societat, especialment en camps específics com la medicina o les finances, cal transparència i interpretabilitat. La majoria dels esforços que s’han fet fins ara s’han centrat a explicar l’aprenentatge supervisat (supervised learning). Aquest projecte pretén utilitzar alguns d’aquests existosos algoritmes d’explicabilitat i aplicar-los a l’aprenentatge per reforç (Reinforcement Learning, RL). Per fer-ho, expliquem les accions d’un agent de RL que juga al joc Breakout d’Atari utilitzant dos algoritmes diferents: explicacions additives de Shapley (SHAP) i explicacions model-agnòstiques localment interpretables (LIME). Hem implementat amb èxit tots dos algoritmes, que produeixen explicacions creïbles i interessants de la mecànica de l’agent. Tanmateix, creiem que la presentació final dels resultats no és òptima per a l’usuari final, ja que no és intuïtiva a primera vista.
- Published
- 2022
27. Justifying Short-Term Load Forecasts Obtained with the Use of Neural Models
- Author
-
Michal K. Grzeszczyk and Tadeusz A. Grzeszczyk
- Subjects
time-series forecasting ,short-term load forecasting ,energy forecasting model ,neural networks ,explainability ,local interpretable model-agnostic explanations ,Control and Optimization ,Renewable Energy, Sustainability and the Environment ,Energy Engineering and Power Technology ,Electrical and Electronic Engineering ,Engineering (miscellaneous) ,Energy (miscellaneous) - Abstract
There is a lot of research on the neural models used for short-term load forecasting (STLF), which is crucial for improving the sustainable operation of energy systems with increasing technical, economic, and environmental requirements. Neural networks are computationally powerful; however, the lack of clear, readable and trustworthy justification of STLF obtained using such models is a serious problem that needs to be tackled. The article proposes an approach based on the local interpretable model-agnostic explanations (LIME) method that supports reliable premises justifying and explaining the forecasts. The use of the proposed approach makes it possible to improve the reliability of heuristic and experimental neural modeling processes, the results of which are difficult to interpret. Explaining the forecasting may facilitate the justification of the selection and the improvement of neural models for STLF, while contributing to a better understanding of the obtained results and broadening the knowledge and experience supporting the enhancement of energy systems security based on reliable forecasts and simplifying dispatch decisions.
- Published
- 2022
- Full Text
- View/download PDF
28. Extending machine learning prediction capabilities by explainable AI in financial time series prediction.
- Author
-
Çeli̇k, Taha Buğra, İcan, Özgür, and Bulut, Elif
- Subjects
HILBERT-Huang transform ,TIME series analysis ,MACHINE learning ,ARTIFICIAL intelligence ,PREDICTION models ,MARKET prices - Abstract
Prediction with higher accuracy is vital for stock market prediction. Recently, considerable amount of effort has been poured into employing machine learning (ML) techniques for successfully predicting stock market price direction. No matter how successful the proposed prediction model is, it can be argued that there occur two major drawbacks for further increasing the prediction accuracy. The first one can be referred as the black box nature of ML techniques, in other words inference from the predictions cannot be explained. Furthermore, due to the complex characteristics of the predicted time series, no matter how sophisticated techniques are employed, it would be very difficult to achieve a marginal increase in accuracy that would meaningfully offset the additional computational burden it brings in. For these two reasons, instead of chasing incremental improvements in accuracy, we propose utilizing an "e X plainable A rtificial I ntelligence" (XAI) approach which can be employed for assessing the reliability of the predictions hence allowing decision maker to abstain from poor decisions which are responsible for declining overall prediction performance. If there would be a measure of how sure the prediction model is on any prediction, the predictions with a relatively higher reliability could be used to make a decision while lower quality decisions could be avoided. In this study, a novel two-stage stacking ensemble model for stock market direction prediction based on ML, empirical mode decomposition (EMD) and XAI is proposed. Our experiments have shown that, proposed prediction model supported with local interpretable model-agnostic explanations (LIME) achieved the highest accuracy of 0.9913 when only the most trusted predictions have been considered on KOSPI dataset and analogous successful results have been obtained from five other major stock market indices. • Accuracy increase in stock market prediction model when cannot be increased further with explainable machine learning. • Novel application method for model explainability with machine learning prediction models. • Performance increase for Empirical Mode Decomposition based prediction models with ensemble learning. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
29. Ensemble-based genetic algorithm explainer with automized image segmentation: A case study on melanoma detection dataset.
- Author
-
Nematzadeh H, García-Nieto J, Navas-Delgado I, and Aldana-Montes JF
- Subjects
- Humans, Oxides, Artificial Intelligence, Melanoma
- Abstract
Explainable Artificial Intelligence (XAI) makes AI understandable to the human user particularly when the model is complex and opaque. Local Interpretable Model-agnostic Explanations (LIME) has an image explainer package that is used to explain deep learning models. The image explainer of LIME needs some parameters to be manually tuned by the expert in advance, including the number of top features to be seen and the number of superpixels in the segmented input image. This parameter tuning is a time-consuming task. Hence, with the aim of developing an image explainer that automizes image segmentation, this paper proposes Ensemble-based Genetic Algorithm Explainer (EGAE) for melanoma cancer detection that automatically detects and presents the informative sections of the image to the user. EGAE has three phases. First, the sparsity of chromosomes in GAs is determined heuristically. Then, multiple GAs are executed consecutively. However, the difference between these GAs are in different number of superpixels in the input image that result in different chromosome lengths. Finally, the results of GAs are ensembled using consensus and majority votings. This paper also introduces how Euclidean distance can be used to calculate the distance between the actual explanation (delineated by experts) and the calculated explanation (computed by the explainer) for accuracy measurement. Experimental results on a melanoma dataset show that EGAE automatically detects informative lesions, and it also improves the accuracy of explanation in comparison with LIME efficiently. The python codes for EGAE, the ground truths delineated by clinicians, and the melanoma detection dataset are available at https://github.com/KhaosResearch/EGAE., Competing Interests: Declaration of Competing Interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (Copyright © 2023 The Author(s). Published by Elsevier Ltd.. All rights reserved.)
- Published
- 2023
- Full Text
- View/download PDF
30. Analysis of input parameters for deep learning-based load prediction for office buildings in different climate zones using eXplainable Artificial Intelligence.
- Author
-
Chung, Woong June and Liu, Chunde
- Subjects
- *
DEEP learning , *ARTIFICIAL intelligence , *OFFICE buildings , *LATIN hypercube sampling , *FORECASTING - Abstract
• Two thousand office building simulation models were generated using Latin hypercube sampling for various climate zones. • The XAI techniques, i.e., SHAP and LIME could identify the essential input variables more effectively compared with SRC. • The SHAP method appears to be superior over LIME method in identifying the most essential input variables. • The types and essential number of input parameters for the deep learning-based model varied with different climate zones. Simplification of input variables can increase the applicability of Artificial Intelligence (AI) in building load prediction. The most essential inputs for AI therefore need to be identified via a significance level test. In this study, the significance of the input parameters was evaluated using the standardized regression coefficient (SRC) and Explainable AI methods, i.e., Local Interpretable Model-agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP). To consider various types of office buildings operating under different climates, the constraints of the U.S. Department of Energy reference buildings with the Latin hypercube sampling were used to generate two thousand building models for each of five climate zones classified by the Köppen and Geiger system. The number and types of essential inputs for deep learning-based building load prediction were identified based on SRC, LIME, and SHAP. By comparison, the SHAP method gave the smallest number of essential inputs for accurate load prediction. In addition, the types of essential inputs varied with different climates. This study presented the potential elimination of weather sensors with time variables depending on the climate conditions. It would help building practitioners and non-experts to determine the essential inputs to build up a simple but accurate deep learning-based building system. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
31. Explainable AI techniques for sepsis diagnosis : Evaluating LIME and SHAP through a user study
- Author
-
Norrie, Christian and Norrie, Christian
- Abstract
Articial intelligence has had a large impact on many industries and transformed some domains quite radically. There is tremendous potential in applying AI to the eld of medical diagnostics. A major issue with applying these techniques to some domains is an inability for AI models to provide an explanation or justication for their predictions. This creates a problem wherein a user may not trust an AI prediction, or there are legal requirements for justifying decisions that are not met. This thesis overviews how two explainable AI techniques (Shapley Additive Explanations and Local Interpretable Model-Agnostic Explanations) can establish a degree of trust for the user in the medical diagnostics eld. These techniques are evaluated through a user study. User study results suggest that supplementing classications or predictions with a post-hoc visualization increases interpretability by a small margin. Further investigation and research utilizing a user study surveyor interview is suggested to increase interpretability and explainability of machine learning results.
- Published
- 2021
32. Comprehensible Machine-Learning-Based Models for the Pre-Emptive Diagnosis of Multiple Sclerosis Using Clinical Data: A Retrospective Study in the Eastern Province of Saudi Arabia.
- Author
-
Olatunji SO, Alsheikh N, Alnajrani L, Alanazy A, Almusairii M, Alshammasi S, Alansari A, Zaghdoud R, Alahmadi A, Basheer Ahmed MI, Ahmed MS, and Alhiyafi J
- Subjects
- Humans, Retrospective Studies, Saudi Arabia, Brain, Machine Learning, Multiple Sclerosis
- Abstract
Multiple Sclerosis (MS) is characterized by chronic deterioration of the nervous system, mainly the brain and the spinal cord. An individual with MS develops the condition when the immune system begins attacking nerve fibers and the myelin sheathing that covers them, affecting the communication between the brain and the rest of the body and eventually causing permanent damage to the nerve. Patients with MS (pwMS) might experience different symptoms depending on which nerve was damaged and how much damage it has sustained. Currently, there is no cure for MS; however, there are clinical guidelines that help control the disease and its accompanying symptoms. Additionally, no specific laboratory biomarker can precisely identify the presence of MS, leaving specialists with a differential diagnosis that relies on ruling out other possible diseases with similar symptoms. Since the emergence of Machine Learning (ML) in the healthcare industry, it has become an effective tool for uncovering hidden patterns that aid in diagnosing several ailments. Several studies have been conducted to diagnose MS using ML and Deep Learning (DL) models trained using MRI images, achieving promising results. However, complex and expensive diagnostic tools are needed to collect and examine imaging data. Thus, the intention of this study is to implement a cost-effective, clinical data-driven model that is capable of diagnosing pwMS. The dataset was obtained from King Fahad Specialty Hospital (KFSH) in Dammam, Saudi Arabia. Several ML algorithms were compared, namely Support Vector Machine (SVM), Decision Tree (DT), Logistic Regression (LR), Random Forest (RF), Extreme Gradient Boosting (XGBoost), Adaptive Boosting (AdaBoost), and Extra Trees (ET). The results indicated that the ET model outpaced the rest with an accuracy of 94.74%, recall of 97.26%, and precision of 94.67%.
- Published
- 2023
- Full Text
- View/download PDF
33. Interpreting Multivariate Time Series for an Organization Health Platform
- Author
-
Saluja, Rohit and Saluja, Rohit
- Abstract
Machine learning-based systems are rapidly becoming popular because it has been realized that machines are more efficient and effective than humans at performing certain tasks. Although machine learning algorithms are extremely popular, they are also very literal and undeviating. This has led to a huge research surge in the field of interpretability in machine learning to ensure that machine learning models are reliable, fair, and can be held liable for their decision-making process. Moreover, in most real-world problems just making predictions using machine learning algorithms only solves the problem partially. Time series is one of the most popular and important data types because of its dominant presence in the fields of business, economics, and engineering. Despite this, interpretability in time series is still relatively unexplored as compared to tabular, text, and image data. With the growing research in the field of interpretability in machine learning, there is also a pressing need to be able to quantify the quality of explanations produced after interpreting machine learning models. Due to this reason, evaluation of interpretability is extremely important. The evaluation of interpretability for models built on time series seems completely unexplored in research circles. This thesis work focused on achieving and evaluating model agnostic interpretability in a time series forecasting problem. The use case discussed in this thesis work focused on finding a solution to a problem faced by a digital consultancy company. The digital consultancy wants to take a data-driven approach to understand the effect of various sales related activities in the company on the sales deals closed by the company. The solution involved framing the problem as a time series forecasting problem to predict the sales deals and interpreting the underlying forecasting model. The interpretability was achieved using two novel model agnostic interpretability techniques, Local interpretable, Maskininlärningsbaserade system blir snabbt populära eftersom man har insett att maskiner är effektivare än människor när det gäller att utföra vissa uppgifter. Även om maskininlärningsalgoritmer är extremt populära, är de också mycket bokstavliga. Detta har lett till en enorm forskningsökning inom området tolkbarhet i maskininlärning för att säkerställa att maskininlärningsmodeller är tillförlitliga, rättvisa och kan hållas ansvariga för deras beslutsprocess. Dessutom löser problemet i de flesta verkliga problem bara att göra förutsägelser med maskininlärningsalgoritmer bara delvis. Tidsserier är en av de mest populära och viktiga datatyperna på grund av dess dominerande närvaro inom affärsverksamhet, ekonomi och teknik. Trots detta är tolkningsförmågan i tidsserier fortfarande relativt outforskad jämfört med tabell-, text- och bilddata. Med den växande forskningen inom området tolkbarhet inom maskininlärning finns det också ett stort behov av att kunna kvantifiera kvaliteten på förklaringar som produceras efter tolkning av maskininlärningsmodeller. Av denna anledning är utvärdering av tolkbarhet extremt viktig. Utvärderingen av tolkbarhet för modeller som bygger på tidsserier verkar helt outforskad i forskarkretsar. Detta uppsatsarbete fokuserar på att uppnå och utvärdera agnostisk modelltolkbarhet i ett tidsserieprognosproblem. Fokus ligger i att hitta lösningen på ett problem som ett digitalt konsultföretag står inför som användningsfall. Det digitala konsultföretaget vill använda en datadriven metod för att förstå effekten av olika försäljningsrelaterade aktiviteter i företaget på de försäljningsavtal som företaget stänger. Lösningen innebar att inrama problemet som ett tidsserieprognosproblem för att förutsäga försäljningsavtalen och tolka den underliggande prognosmodellen. Tolkningsförmågan uppnåddes med hjälp av två nya tekniker för agnostisk tolkbarhet, lokala tolkbara modellagnostiska förklaringar (LIME) och Shapley additiva förklaringar (SHAP). Förklarin
- Published
- 2020
34. Justifying Short-Term Load Forecasts Obtained with the Use of Neural Models.
- Author
-
Grzeszczyk, Tadeusz A. and Grzeszczyk, Michal K.
- Subjects
- *
LOAD forecasting (Electric power systems) , *FORECASTING , *ENERGY security , *SECURITY systems - Abstract
There is a lot of research on the neural models used for short-term load forecasting (STLF), which is crucial for improving the sustainable operation of energy systems with increasing technical, economic, and environmental requirements. Neural networks are computationally powerful; however, the lack of clear, readable and trustworthy justification of STLF obtained using such models is a serious problem that needs to be tackled. The article proposes an approach based on the local interpretable model-agnostic explanations (LIME) method that supports reliable premises justifying and explaining the forecasts. The use of the proposed approach makes it possible to improve the reliability of heuristic and experimental neural modeling processes, the results of which are difficult to interpret. Explaining the forecasting may facilitate the justification of the selection and the improvement of neural models for STLF, while contributing to a better understanding of the obtained results and broadening the knowledge and experience supporting the enhancement of energy systems security based on reliable forecasts and simplifying dispatch decisions. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
35. Energy consumption, flow characteristics and energy-efficient design of cup-shape blade stirred tank reactors: Computational fluid dynamics and artificial neural network investigation.
- Author
-
Zhao, Shuchun, Guo, Junheng, Dang, Xiuhu, Ai, Bingyan, Zhang, Minqing, Li, Wei, and Zhang, Jinli
- Subjects
- *
ARTIFICIAL neural networks , *COMPUTATIONAL fluid dynamics , *ENERGY consumption , *BACK propagation , *VECTOR quantization , *THERMAL hydraulics - Abstract
Developing high-efficient and green energy-saving stirred tank reactors is of vital importance in liquid-liquid processing industries. Therefore, this work revealed the energy consumption, flow characteristics, and explored the energy-saving strategies of novel-designed Cup-shape Blade (CB) stirred tank reactors characterized by easy fabrication, low energy consumption and large pumping flow rate. Computational Fluid Dynamics (CFD) and Artificial Neural Network (ANN) were conducted to investigate the influences of geometrical parameters on liquid-liquid power number (N P), flow number (N Q) and flow patterns. The improved method for hyper-parameter specification indicates that the Bayesian regulation training algorithm and the two-hidden-layer scheme possess the best performance. Back Propagation Neural Network (BPNN) presents the best predictive accuracy for N P and N Q , and Learning Vector Quantization (LVQ) network shows the best classification capability of flow patterns. Analyses based on the ANN modeling and the "Local Interpretable Model-agnostic Explanations" method show the joint and complicated effects of geometrical parameters on energy consumption and flow characteristics. Procedures were proposed to ascertain the energy-efficiency blade configuration based on ANN. [Display omitted] • CFD and ANN method were used to study the performance of CB stirred tanks. • An optimized method was adopted to ascertain the details of ANN. • Local sensitivities of input variables were implemented. • The procedures to select an energy-efficient blade configuration were proposed. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
36. Implementing Machine Learning in the Credit Process of a Learning Organization While Maintaining Transparency Using LIME
- Author
-
Malmberg, Jacob, Nystad Öhman, Marcus, Hotti, Alexandra, Malmberg, Jacob, Nystad Öhman, Marcus, and Hotti, Alexandra
- Abstract
To determine whether a credit limit for a corporate client should be changed, a financial institution writes a PM containingtext and financial data that then is assessed by a credit committee which decides whether to increase the limit or not. To make thisprocess more efficient, machine learning algorithms was used to classify the credit PMs instead of a committee. Since most machinelearning algorithms are black boxes, the LIME framework was used to find the most important features driving the classification. Theresults of this study show that credit memos can be classified with high accuracy and that LIME can be used to indicate which parts ofthe memo had the biggest impact. This implicates that the credit process could be improved by utilizing machine learning, whilemaintaining transparency. However, machine learning may disrupt learning processes within the organization., För att bedöma om en kreditlimit för ett företag ska förändras eller inte skriver ett finansiellt institut ett PM innehållande text och finansiella data. Detta PM granskas sedan av en kreditkommitté som beslutar om limiten ska förändras eller inte. För att effektivisera denna process användes i denna rapport maskininlärning istället för en kreditkommitté för att besluta om limiten ska förändras. Eftersom de flesta maskininlärningsalgoritmer är svarta lådor så användes LIME-ramverket för att hitta de viktigaste drivarna bakom klassificeringen. Denna studies resultat visar att kredit-PM kan klassificeras med hög noggrannhet och att LIME kan visa vilken del av ett PM som hade störst påverkan vid klassificeringen. Implikationerna av detta är att kreditprocessen kan förbättras av maskininlärning, utan att förlora transparens. Maskininlärning kan emellertid störa lärandeprocesser i organisationen, varför införandet av dessa algoritmer bör vägas mot hur betydelsefullt det är att bevara och utveckla kunskap inom organisationen.
- Published
- 2018
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.