147 results
Search Results
2. Challenges and strategies for wide-scale artificial intelligence (AI) deployment in healthcare practices: A perspective for healthcare organizations.
- Author
-
Esmaeilzadeh, Pouyan
- Abstract
Healthcare organizations have realized that Artificial intelligence (AI) can provide a competitive edge through personalized patient experiences, improved patient outcomes, early diagnosis, augmented clinician capabilities, enhanced operational efficiencies, or improved medical service accessibility. However, deploying AI-driven tools in the healthcare ecosystem could be challenging. This paper categorizes AI applications in healthcare and comprehensively examines the challenges associated with deploying AI in medical practices at scale. As AI continues to make strides in healthcare, its integration presents various challenges, including production timelines, trust generation, privacy concerns, algorithmic biases, and data scarcity. The paper highlights that flawed business models and wrong workflows in healthcare practices cannot be rectified merely by deploying AI-driven tools. Healthcare organizations should re-evaluate root problems such as misaligned financial incentives (e.g., fee-for-service models), dysfunctional medical workflows (e.g., high rates of patient readmissions), poor care coordination between different providers, fragmented electronic health records systems, and inadequate patient education and engagement models in tandem with AI adoption. This study also explores the need for a cultural shift in viewing AI not as a threat but as an enabler that can enhance healthcare delivery and create new employment opportunities while emphasizing the importance of addressing underlying operational issues. The necessity of investments beyond finance is discussed, emphasizing the importance of human capital, continuous learning, and a supportive environment for AI integration. The paper also highlights the crucial role of clear regulations in building trust, ensuring safety, and guiding the ethical use of AI, calling for coherent frameworks addressing transparency, model accuracy, data quality control, liability, and ethics. Furthermore, this paper underscores the importance of advancing AI literacy within academia to prepare future healthcare professionals for an AI-driven landscape. Through careful navigation and proactive measures addressing these challenges, the healthcare community can harness AI's transformative power responsibly and effectively, revolutionizing healthcare delivery and patient care. The paper concludes with a vision and strategic suggestions for the future of healthcare with AI, emphasizing thoughtful, responsible, and innovative engagement as the pathway to realizing its full potential to unlock immense benefits for healthcare organizations, physicians, nurses, and patients while proactively mitigating risks. • This study categorizes AI applications in healthcare and analyzes their deployment challenges at scale in medical practices. • The paper emphasizes a cultural shift to view AI as a healthcare delivery enhancer and job creator, not a threat. • To integrate AI, investments in finance and human capital, continuous learning, and a supportive environment are needed. • Clear regulatory frameworks should be developed to build trust, ensure safety, and guide the ethical use of AI in healthcare. • The study provides strategic suggestions for responsible AI use in healthcare to maximize benefits and minimize risks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. HR-BGCN [formula omitted] Predicting readmission for heart failure from electronic health records.
- Author
-
Ma, Huiting, Li, Dengao, Zhao, Jumin, Li, Wenjing, Fu, Jian, and Li, Chunxia
- Abstract
Heart failure has become a huge public health problem, and failure to accurately predict readmission will further lead to the disease's high cost and high mortality. The construction of readmission prediction model can assist doctors in making decisions to prevent patients from deteriorating and reduce the cost burden. This paper extracts the patient discharge records from the MIMIC-III database. It divides the patients into three research categories: no readmission, readmission within 30 days, and readmission after 30 days, to predict the readmission of patients. We propose the HR-BGCN model to predict the readmission of patients. First, we use the Adaptive-TMix to improve the prediction indicators of a few categories and reduce the impact of unbalanced categories. Then, the knowledge-informed graph attention mechanism is proposed. By introducing a document-level explicit diagram structure, the coding ability of graph node features is significantly improved. The paragraph-level representation obtained through graph learning is combined with the context token-level representation of BERT, and finally, the multi-classification task is carried out. We also compare several typical graph learning classification models to verify the model's effectiveness, such as the IA-GCN model, GAT model, etc. The results show that the average F1 score of the HR-BGCN model proposed in this paper for 30-day readmission of heart failure patients is 88.26%, and the average accuracy is 90.47%. The HR-BGCN model is significantly better than the graph learning classification model for predicting heart failure readmission. It can help doctors predict the 30-day readmission of patients, then reduce the readmission rate of patients. • The clinical notes (unstructured data) are used for prognostic prediction. • The Adaptive-TMix is proposed to reduce the impact of unbalanced categories. • The proposed knowledge-informed graph attention mechanism can improve coding ability of graph node features. • The model can complete a multi-class prediction task and perform well. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. ECG-based cardiac arrhythmias detection through ensemble learning and fusion of deep spatial–temporal and long-range dependency features.
- Author
-
Din, Sadia, Qaraqe, Marwa, Mourad, Omar, Qaraqe, Khalid, and Serpedin, Erchin
- Abstract
Cardiac arrhythmia is one of the prime reasons for death globally. Early diagnosis of heart arrhythmia is crucial to provide timely medical treatment. Heart arrhythmias are diagnosed by analyzing the electrocardiogram (ECG) of patients. Manual analysis of ECG is time-consuming and challenging. Hence, effective automated detection of heart arrhythmias is important to produce reliable results. Different deep-learning techniques to detect heart arrhythmias such as Convolutional Neural Network (CNN), Long Short-Term Memory (LSTM), Transformer, and Hybrid CNN–LSTM were proposed. However, these techniques, when used individually, are not sufficient to effectively learn multiple features from the ECG signal. The fusion of CNN and LSTM overcomes the limitations of CNN in the existing studies as CNN–LSTM hybrids can extract spatiotemporal features. However, LSTMs suffer from long-range dependency issues due to which certain features may be ignored. Hence, to compensate for the drawbacks of the existing models, this paper proposes a more comprehensive feature fusion technique by merging CNN, LSTM, and Transformer models. The fusion of these models facilitates learning spatial, temporal, and long-range dependency features, hence, helping to capture different attributes of the ECG signal. These features are subsequently passed to a majority voting classifier equipped with three traditional base learners. The traditional learners are enriched with deep features instead of handcrafted features. Experiments are performed on the MIT-BIH arrhythmias database and the model performance is compared with that of the state-of-art models. Results reveal that the proposed model performs better than the existing models yielding an accuracy of 99.56%. • This paper proposes a rich feature fusion technique for cardiac arrhythmia detection from ECG by merging three deep learning models, namely, CNN, CNN–LSTM, and Transformer. • Integrating these models combines the advantages of all these individual models, overcomes their limitations, and learns significant features from the input ECG signal. • The deep spatial, temporal, and long-range dependency patterns learned by the aforementioned models are fused via concatenation and fed to a majority voting classifier with three traditional base learners: SVM, LR, and RF. • MIT-BIH Arrhythmia database is used for experimentation. A comparison of the results with the baseline models and state-of-the-art models shows that the proposed model outdoes the existing models achieving an accuracy of 95.56% and an F-score of 99.34%. • The results show that the proposed model outperforms the existing models in terms of precision, recall, F-score, and accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. The Concordance Index decomposition: A measure for a deeper understanding of survival prediction models.
- Author
-
Alabdallah, Abdallah, Ohlsson, Mattias, Pashami, Sepideh, and Rögnvaldsson, Thorsteinn
- Abstract
The Concordance Index (C-index) is a commonly used metric in Survival Analysis for evaluating the performance of a prediction model. In this paper, we propose a decomposition of the C-index into a weighted harmonic mean of two quantities: one for ranking observed events versus other observed events, and the other for ranking observed events versus censored cases. This decomposition enables a finer-grained analysis of the relative strengths and weaknesses between different survival prediction methods. The usefulness of this decomposition is demonstrated through benchmark comparisons against classical models and state-of-the-art methods, together with the new variational generative neural-network-based method (SurVED) proposed in this paper. The performance of the models is assessed using four publicly available datasets with varying levels of censoring. Using the C-index decomposition and synthetic censoring, the analysis shows that deep learning models utilize the observed events more effectively than other models. This allows them to keep a stable C-index in different censoring levels. In contrast to such deep learning methods, classical machine learning models deteriorate when the censoring level decreases due to their inability to improve on ranking the events versus other events. • The C-Index Decomposition is proposed as a weighted harmonic mean of the C-index of the events vs other events and the C-index of events vs censored cases. • The C-Index Decomposition gives a more detailed image of the survival model performance. • SurVED; a new survival model based on Variational inference formulation was shown to be better at improving performance with more observed event. • Models which seem to perform similarly using the C-index, showed different behavior with respect to the terms of the C-index decomposition. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Diseases diagnosis based on artificial intelligence and ensemble classification.
- Author
-
Rabie, Asmaa H. and Saleh, Ahmed I.
- Abstract
In recent years, Computer Aided Diagnosis (CAD) has become an important research area that attracted a lot of researchers. In medical diagnostic systems, several attempts have been made to build and enhance CAD applications to avoid errors that can cause dangerously misleading medical treatments. The most exciting opportunity for promoting the performance of CAD system can be accomplished by integrating Artificial Intelligence (AI) in medicine. This allows the effective automation of traditional manual workflow, which is slow, inaccurate and affected by human errors. This paper aims to provide a complete Computer Aided Disease Diagnosis (CAD2) strategy based on Machine Learning (ML) techniques that can help clinicians to make better medical decisions. The proposed CAD2 consists of three main sequential phases, namely; (i) Outlier Rejection Phase (ORP), (ii) Feature Selection Phase (FSP), and (iii) Classification Phase (CP). ORP is implemented to reject outliers using new Outlier Rejection Technique (ORT) that contains two sequential stages called Fast Outlier Rejection (FOR) and Accurate Outlier Rejection (AOR). The most informative features are selected through FSP using Hybrid Selection Technique (HST). HST includes two main stages called Quick Selection Stage (QS2) using fisher score as a filter method and Precise Selection Stage (PS2) using a Hybrid Bio-inspired Optimization (HBO) technique as a wrapper method. Finally, actual diagnose takes place through CP, which relies on Ensemble Classification Technique (ECT). The proposed CAD2 has been tested experimentally against recent disease diagnostic strategies using two different datasets in which the first contains several diseases, while the second includes data for Covid-19 patients only. Experimental results have proven the high efficiency of the proposed CAD2 in terms of accuracy, error, precision, and recall compared with other competitors. Additionally, CAD2 strategy provides the best Wilcoxon signed rank test and Friedman test measurements against other strategies according to both datasets. It is concluded that CAD2 strategy based on ORP, FSP, and CP gave an accurate diagnosis compared to other strategies because it gave the highest accuracy and the lowest error and implementation time. • This paper provides Computer Aided Disease Diagnosis (CAD2) strategy by AI methods. • CAD2 includes outlier rejection, feature selection, and classification phases. • Rejection phase eliminates outliers but selection phase selects the best features. • Actual diagnosis takes place through classification phase using AI techniques. • The provided CAD2 introduces the maximum diagnose accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. BGRL: Basal Ganglia inspired Reinforcement Learning based framework for deep brain stimulators.
- Author
-
Agarwal, Harsh and Rathore, Heena
- Abstract
Deep Brain Stimulation (DBS) is an implantable medical device used for electrical stimulation to treat neurological disorders. Traditional DBS devices provide fixed frequency pulses, but personalized adjustment of stimulation parameters is crucial for optimal treatment. This paper introduces a Basal Ganglia inspired Reinforcement Learning (BGRL) approach, incorporating a closed-loop feedback mechanism to suppress neural synchrony during neurological fluctuations. The BGRL approach leverages the resemblance between the Basal Ganglia region of brain by incorporating the actor–critic architecture of reinforcement learning (RL). Simulation results demonstrate that BGRL significantly reduces synchronous electrical pulses compared to other standard RL algorithms. BGRL algorithm outperforms existing RL methods in terms of suppression capability and energy consumption, validated through comparisons using ensemble oscillators. Results shown in the paper demonstrate BGRL suppressed the synchronous electrical pulses across three signaling regimes namely regular, chaotic and bursting by 40%, 146% and 40% respectively as compared to soft actor–critic model. BGRL shows promise in effectively suppressing neural synchrony in DBS therapy, providing an efficient alternative to open-loop methodologies. • Stabilize the synchronization and convergence points allowing for more accurate estimation. • Avoid overestimated bias to generalize the model well. • Adds noise to the target action to exploit the Q-function errors by smoothing out Q-values. • Ran simulations which capture the characteristics of closed loop DBS. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Value function assessment to different RL algorithms for heparin treatment policy of patients with sepsis in ICU.
- Author
-
Liu, Jiang, Xie, Yihao, Shu, Xin, Chen, Yuwen, Sun, Yizhu, Zhong, Kunhua, Liang, Hao, Li, Yujie, Yang, Chunyong, Han, Yan, Zou, Yuwei, Zhuyi, Ziting, Huang, Jiahao, Li, Junhong, Hu, Xiaoyan, and Yi, Bin
- Abstract
Heparin is a critical aspect of managing sepsis after abdominal surgery, which can improve microcirculation, protect organ function, and reduce mortality. However, there is no clinical evidence to support decision-making for heparin dosage. This paper proposes a model called SOFA-MDP, which utilizes SOFA scores as states of MDP, to investigate clinic policies. Different algorithms provide different value functions, making it challenging to determine which value function is more reliable. Due to ethical restrictions, we cannot test all policies on patients. To address this issue, we proposed two value function assessment methods: action similarity rate and relative gain. We experimented with heparin treatment policies for sepsis patients after abdominal surgery using MIMIC-IV. In the experiments, TD (0) shows the most reliable performance. Using the action similarity rate and relative gain to assess AI policy from TD (0) , the agreement rates between AI policy and "good" physician's actual treatment are 64.6% and 73.2%, while the agreement rates between AI policy and "bad" physician's actual treatment are 44.1% and 35.8%, the gaps are 20.5% and 37.4%, respectively. External validation using action similarity rate and relative gain based on eICU resulted in agreement rates of 61.5% and 69.1% with the "good" physician's treatment, and 45.2% and 38.3% with the "bad" physician's treatment, with gaps of 16.3% and 30.8%, respectively. In conclusion, the model provides instructive support for clinical decisions, and the evaluation methods accurately distinguish reliable and unreasonable outcomes. • This paper proposes a model called SOFA-MDP , which utilizes SOFA scores as states and the difference of SOFA scores of successive states as reward function, to investigate heparin treatment policy of patients with sepsis in ICU. • Moreover, we present two novel assessment methods: action similarity rate and relative gain , to estimate the clinic policies. The experimental results show that our SOFA-MDP model offers instructive support for clinical decisions of heparin treatment and our assessment methods can distinguish between reliable and unreasonable outcomes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Depression detection for twitter users using sentiment analysis in English and Arabic tweets.
- Author
-
Helmy, AbdelMoniem, Nassar, Radwa, and Ramdan, Nagy
- Abstract
Since depression often results in suicidal thoughts and leaves a person severely disabled daily, there is an elevated risk of premature mortality due to mental problems caused by depression. Therefore, it's crucial to identify the patient's mental illness as soon as possible. People are increasingly using social media platforms to express their opinions and share daily activities, which makes online platforms rich sources of early depression detection. The contribution of this paper is multifold. First , it presents five machine-learning models for Arabic and English depression detection using Twitter text. The best model for Arabic text achieved an f1-score of 96.6 % for binary classification to depressed and Non_dep. For English text without negation, the model achieved 92 % for binary classification and 88 % for multi-classification (depressed, indifferent, happy). For English text with negation, an 87 %, and 85 % f1 score was achieved for binary and multi-classification respectively. Second , the work introduced a manually annotated Arabic_Dep_tweets_10,000 corpus of 10.000 Arabic tweets, which covered neutral tweets as well as a variety of depressed and happy terms. In addition, two automatically annotated English corpora, Eng_without_negation_60.000 corpus of 60,172 English tweets and Eng_with_negation_57.000 corpus of 57,392 English tweets. Both covered a wide range of depressed and cheerful terms; however, Negation was included in the Eng_with_negation_57.000 corpus. Finally , this paper exposes a depression-detection web application which implements our optimal models to detect tweets that contain depression symptoms and predict depression trends for a person either using English or Arabic language. • More than two-thirds of suicides each year are caused by depression, which is the most common mental illness. • Social media posts can be a useful tool for tracking a variety of mental health conditions, including depression. • The lack of adequate Arabic corpora makes it difficult to use machine learning techniques in depression detection for the Arabic people. • Existing models for English language use binary classification only and does not handle negation in the text. • Twittpy is a web application developed to detect depression symptoms for persons using their text from Arabic or English tweets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Improving diagnosis and outcome prediction of gastric cancer via multimodal learning using whole slide pathological images and gene expression.
- Author
-
Xie, Yuzhang, Sang, Qingqing, Da, Qian, Niu, Guoshuai, Deng, Shijie, Feng, Haoran, Chen, Yunqin, Li, Yuan-Yuan, Liu, Bingya, Yang, Yang, and Dai, Wentao
- Abstract
For the diagnosis and outcome prediction of gastric cancer (GC), machine learning methods based on whole slide pathological images (WSIs) have shown promising performance and reduced the cost of manual analysis. Nevertheless, accurate prediction of GC outcome may rely on multiple modalities with complementary information, particularly gene expression data. Thus, there is a need to develop multimodal learning methods to enhance prediction performance. In this paper, we collect a dataset from Ruijin Hospital and propose a multimodal learning method for GC diagnosis and outcome prediction, called GaCaMML, which is featured by a cross-modal attention mechanism and Per-Slide training scheme. Additionally, we perform feature attribution analysis via integrated gradient (IG) to identify important input features. The proposed method improves prediction accuracy over the single-modal learning method on three tasks, i.e., survival prediction (by 4.9% on C-index), pathological stage classification (by 11.6% on accuracy), and lymph node classification (by 12.0% on accuracy). Especially, the Per-Slide strategy addresses the issue of a high WSI-to-patient ratio and leads to much better results compared with the Per-Person training scheme. For the interpretable analysis, we find that although WSIs dominate the prediction for most samples, there is still a substantial portion of samples whose prediction highly relies on gene expression information. This study demonstrates the great potential of multimodal learning in GC-related prediction tasks and investigates the contribution of WSIs and gene expression, respectively, which not only shows how the model makes a decision but also provides insights into the association between macroscopic pathological phenotypes and microscopic molecular features. [Display omitted] • GaCaMML outperforms the single-modal learning method across three tasks on accuracy. • Per-Slide strategy enhances prediction results by optimizing WSI-to-patient ratio. • WSI features prevail in gastric cancer outcome prediction for most samples. • Certain genes like MMPs and TP53 significantly impact prediction results. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. MICIL: Multiple-Instance Class-Incremental Learning for skin cancer whole slide images.
- Author
-
Meseguer, Pablo, del Amor, Rocío, and Naranjo, Valery
- Abstract
Artificial intelligence (AI) agents encounter the problem of catastrophic forgetting when they are trained in sequentially with new data batches. This issue poses a barrier to the implementation of AI-based models in tasks that involve ongoing evolution, such as cancer prediction. Moreover, whole slide images (WSI) play a crucial role in cancer management, and their automated analysis has become increasingly popular in assisting pathologists during the diagnosis process. Incremental learning (IL) techniques aim to develop algorithms capable of retaining previously acquired information while also acquiring new insights to predict future data. Deep IL techniques need to address the challenges posed by the gigapixel scale of WSIs, which often necessitates the use of multiple instance learning (MIL) frameworks. In this paper, we introduce an IL algorithm tailored for analyzing WSIs within a MIL paradigm. The proposed M ultiple I nstance C lass- I ncremental L earning (MICIL) algorithm combines MIL with class-IL for the first time, allowing for the incremental prediction of multiple skin cancer subtypes from WSIs within a class-IL scenario. Our framework incorporates knowledge distillation and data rehearsal, along with a novel embedding-level distillation, aiming to preserve the latent space at the aggregated WSI level. Results demonstrate the algorithm's effectiveness in addressing the challenge of balancing IL-specific metrics, such as intransigence and forgetting, and solving the plasticity-stability dilemma. • An incremental learning algorithm is applied under the MIL paradigm to analyze WSI. • A public dataset featuring patch-level embeddings of multiple spindle-cell neoplasms. • A WSI embedding-level distillation to prevent shift in the aggregated latent space. • Comprehensive experiments address the plasticity-stability dilemma in MIL frameworks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Agent-based approaches for biological modeling in oncology: A literature review.
- Author
-
Stephan, Simon, Galland, Stéphane, Labbani Narsis, Ouassila, Shoji, Kenji, Vachenc, Sébastien, Gerart, Stéphane, and Nicolle, Christophe
- Abstract
Computational modeling involves the use of computer simulations and models to study and understand real-world phenomena. Its application is particularly relevant in the study of potential interactions between biological elements. It is a promising approach to understand complex biological processes and predict their behavior under various conditions. This paper is a review of the recent literature on computational modeling of biological systems. Our study focuses on the field of oncology and the use of artificial intelligence (AI) and, in particular, agent-based modeling (ABM), between 2010 and May 2023. Most of the articles studied focus on improving the diagnosis and understanding the behaviors of biological entities, with metaheuristic algorithms being the models most used. Several challenges are highlighted regarding increasing and structuring knowledge about biological systems, developing holistic models that capture multiple scales and levels of organization, reproducing emergent behaviors of biological systems, validating models with experimental data, improving computational performance of models and algorithms, and ensuring privacy and personal data protection are discussed. • Use of multi-agent systems to study biological interactions in oncology. • Multi-agent systems enable modeling of cell signaling network dynamics. • Review of multi-agent systems application for personalized therapeutic strategies. • Meta-heuristic algorithms found to be the most used models in recent studies. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Hematologic cancer diagnosis and classification using machine and deep learning: State-of-the-art techniques and emerging research directives.
- Author
-
Patel, Hema, Shah, Himal, Patel, Gayatri, and Patel, Atul
- Abstract
Hematology is the study of diagnosis and treatment options for blood diseases, including cancer. Cancer is considered one of the deadliest diseases across all age categories. Diagnosing such a deadly disease at the initial stage is essential to cure the disease. Hematologists and pathologists rely on microscopic evaluation of blood or bone marrow smear images to diagnose blood-related ailments. The abundance of overlapping cells, cells of varying densities among platelets, non-illumination levels, and the amount of red and white blood cells make it more difficult to diagnose illness using blood cell images. Pathologists are required to put more effort into the traditional, time-consuming system. Nowadays, it becomes possible with machine learning and deep learning techniques, to automate the diagnostic processes, categorize microscopic blood cells, and improve the accuracy of the procedure and its speed as the models developed using these methods may guide an assisting tool. In this article, we have acquired, analyzed, scrutinized, and finally selected around 57 research papers from various machine learning and deep learning methodologies that have been employed in the diagnosis of leukemia and its classification over the past 20 years, which have been published between the years 2003 and 2023 by PubMed, IEEE, Science Direct, Google Scholar and other pertinent sources. Our primary emphasis is on evaluating the advantages and limitations of analogous research endeavors to provide a concise and valuable research directive that can be of significant utility to fellow researchers in the field. • Use of machine learning techniques for the diagnostic purpose • Diagnosis and classification of leukemia using computer aided methods • Systematic review of the work done so far in the field of leukemia research • Machine learning and deep learning are capable to proceed large amount of data for higher accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Learning the cellular activity representation based on gene regulatory networks for prediction of tumor response to drugs.
- Author
-
Xie, Xinping, Wang, Fengting, Wang, Guanfu, Zhu, Weiwei, Du, Xiaodong, and Wang, Hongqiang
- Abstract
Predicting the response of tumor cells to anti-tumor drugs is critical to realizing cancer precision medicine. Currently, most existing methods ignore the regulatory relationships between genes and thus have unsatisfactory predictive performance. In this paper, we propose to predict anti-tumor drug efficacy via learning the activity representation of tumor cells based on a priori knowledge of gene regulation networks (GRNs). Specifically, the method simulates the cellular biosystem by synthesizing a cell-gene activity network and then infers a new low-dimensional activity representation for tumor cells from the raw high-dimensional expression profile. The simulated cell-gene network mainly comprises known gene regulatory networks collected from multiple resources and fuses tumor cells by linking them to hotspot genes that are over- or under-expressed in them. The resulting activity representation could not only reflect the shallow expression profile (hotspot genes) but also mines in-depth information of gene regulation activity in tumor cells before treatment. Finally, we build deep learning models on the activity representation for predicting drug efficacy in tumor cells. Experimental results on the benchmark GDSC dataset demonstrate the superior performance of the proposed method over SOTA methods with the highest AUC of 0.954 in the efficacy label prediction and the best R2 of 0.834 in the regression of half maximal inhibitory concentration (IC50) values, suggesting the potential value of the proposed method in practice. • Propose a drug efficacy prediction method based on the cellular activity representation • The activity representation learned from a cell-gene network combining GRNs and expression profiles • The predictive ability of the activity representation depends on simulating the real biosystem. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Application of machine learning in affordable and accessible insulin management for type 1 and 2 diabetes: A comprehensive review.
- Author
-
Eghbali-Zarch, Maryam and Masoud, Sara
- Abstract
Proper insulin management is vital for maintaining stable blood sugar levels and preventing complications associated with diabetes. However, the soaring costs of insulin present significant challenges to ensuring affordable management. This paper conducts a comprehensive review of current literature on the application of machine learning (ML) in insulin management for diabetes patients, particularly focusing on enhancing affordability and accessibility within the United States. The review encompasses various facets of insulin management, including dosage calculation and response, prediction of blood glucose and insulin sensitivity, initial insulin estimation, resistance prediction, treatment adherence, complications, hypoglycemia prediction, and lifestyle modifications. Additionally, the study identifies key limitations in the utilization of ML within the insulin management literature and suggests future research directions aimed at furthering accessible and affordable insulin treatments. These proposed directions include exploring insurance coverage, optimizing insulin type selection, assessing the impact of biosimilar insulin and market competition, considering mental health factors, evaluating insulin delivery options, addressing cost-related issues affecting insulin usage and adherence, and selecting appropriate patient cost-sharing programs. By examining the potential of ML in addressing insulin management affordability and accessibility, this work aims to envision improved and cost-effective insulin management practices. It not only highlights existing research gaps but also offers insights into future directions, guiding the development of innovative solutions that have the potential to revolutionize insulin management and benefit patients reliant on this life-saving treatment. • Reviewing literature on application of machine learning for insulin management • Analyzing insulin affordability and accessibility aspects of insulin treatment • Identifying limitations and highlighting research gaps of existing literature • Proposing future research directions for improved affordable insulin care [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Harnessing machine learning for EEG signal analysis: Innovations in depth of anaesthesia assessment.
- Author
-
Schmierer, Thomas, Li, Tianning, and Li, Yan
- Abstract
Anaesthesia, crucial to surgical practice, is undergoing renewed scrutiny due to the integration of artificial intelligence in its medical use. The precise control over the temporary loss of consciousness is vital to ensure safe, pain-free procedures. Traditional methods of depth of anaesthesia (DoA) assessment, reliant on physical characteristics, have proven inconsistent due to individual variations. In response, electroencephalography (EEG) techniques have emerged, with indices such as the Bispectral Index offering quantifiable assessments. This literature review explores the current scope and frontier of DoA research, emphasising methods utilising EEG signals for effective clinical monitoring. This review offers a critical synthesis of recent advances, specifically focusing on electroencephalography (EEG) techniques and their role in enhancing clinical monitoring. By examining 117 high-impact papers, the review delves into the nuances of feature extraction, model building, and algorithm design in EEG-based DoA analysis. Comparative assessments of these studies highlight their methodological approaches and performance, including clinical correlations with established indices like the Bispectral Index. The review identifies knowledge gaps, particularly the need for improved collaboration for data access, which is essential for developing superior machine learning models and real-time predictive algorithms for patient management. It also calls for refined model evaluation processes to ensure robustness across diverse patient demographics and anaesthetic agents. The review underscores the potential of technological advancements to enhance precision, safety, and patient outcomes in anaesthesia, paving the way for a new standard in anaesthetic care. The findings of this review contribute to the ongoing discourse on the application of EEG in anaesthesia, providing insights into the potential for technological advancement in this critical area of medical practice. • Review of the current depth of anaesthesia analysis methods • Emphasis on machine learning techniques based on EEG signal analysis • Trends and future direction of EEG-based depth of anaesthesia analysis methods [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. Pathways to democratized healthcare: Envisioning human-centered AI-as-a-service for customized diagnosis and rehabilitation.
- Author
-
Turchi, Tommaso, Prencipe, Giuseppe, Malizia, Alessio, Filogna, Silvia, Latrofa, Francesco, and Sgandurra, Giuseppina
- Abstract
The ongoing digital revolution in the healthcare sector, emphasized by bodies like the US Food and Drug Administration (FDA), is paving the way for a shift towards person-centric healthcare models. These models consider individual needs, turning patients from passive recipients to active participants. A key factor in this shift is Artificial Intelligence (AI), which has the capacity to revolutionize healthcare delivery due to its ability to personalize it. With the rise of software in healthcare and the proliferation of the Internet of Things (IoT), a surge of digital data is being produced. This data, alongside improvements in AI's explainability, is facilitating the spread of person-centric healthcare models, aiming at improving health management and patient experience. This paper outlines a human-centered methodology for the development of an AI-as-a-service platform with the goal of broadening access to personalized healthcare. This approach places humans at its core, aiming to augment, not replace, human capabilities and integrate in current processes. The primary research question guiding this study is: "How can Human-Centered AI principles be considered when designing an AI-as-a-service platform that democratizes access to personalized healthcare?" This informed both our research direction and investigation. Our approach involves a design fiction methodology, engaging clinicians from different domains to gather their perspectives on how AI can meet their needs by envisioning potential future scenarios and addressing possible ethical and social challenges. Additionally, we incorporate Meta-Design principles, investigating opportunities for users to modify the AI system based on their experiences. This promotes a platform that evolves with the user and considers many different perspectives. • Design fiction method effectively probes AI healthcare scenarios, illuminating challenges and opportunities. • Meta-design opportunities identified in personalizing AI-based healthcare interventions. • User journey maps reveal potential biases in AI decision-making and underscore ethical considerations. • AI integration in healthcare requires collaborative efforts between HCI researchers and clinicians. • Aggregated findings across case studies underscore the need for an AI-as-a-service platform prioritizing ethics and personalization. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Cellular data extraction from multiplexed brain imaging data using self-supervised Dual-loss Adaptive Masked Autoencoder.
- Author
-
Ly, Son T., Lin, Bai, Vo, Hung Q., Maric, Dragan, Roysam, Badrinath, and Nguyen, Hien V.
- Abstract
Reliable large-scale cell detection and segmentation is the fundamental first step to understanding biological processes in the brain. The ability to phenotype cells at scale can accelerate preclinical drug evaluation and system-level brain histology studies. The impressive advances in deep learning offer a practical solution to cell image detection and segmentation. Unfortunately, categorizing cells and delineating their boundaries for training deep networks is an expensive process that requires skilled biologists. This paper presents a novel self-supervised Dual-Loss Adaptive Masked Autoencoder (DAMA) for learning rich features from multiplexed immunofluorescence brain images. DAMA's objective function minimizes the conditional entropy in pixel-level reconstruction and feature-level regression. Unlike existing self-supervised learning methods based on a random image masking strategy, DAMA employs a novel adaptive mask sampling strategy to maximize mutual information and effectively learn brain cell data. To the best of our knowledge, this is the first effort to develop a self-supervised learning method for multiplexed immunofluorescence brain images. Our extensive experiments demonstrate that DAMA features enable superior cell detection, segmentation, and classification performance without requiring many annotations. In addition, to examine the generalizability of DAMA, we also experimented on TissueNet, a multiplexed imaging dataset comprised of two-channel fluorescence images from six distinct tissue types, captured using six different imaging platforms. Our code is publicly available at https://github.com/hula-ai/DAMA. • We present a novel information-theoretic self-supervised method for multiplexed brain image analysis. Our theoretical analysis shows that the proposed objective function maximizes the mutual information between the input image and self-supervised labels. • We design the first adaptive mask sampling strategy for self-supervised learning models. We explain how this adaptive sampling helps the model focus on underrepresented features and improve the representation. • Extensive experiments on cell detection and classification are provided to validate the effectiveness of DAMA. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Intelligent decision support systems for dementia care: A scoping review.
- Author
-
Andargoli, Amirhossein Eslami, Ulapane, Nalika, Nguyen, Tuan Anh, Shuakat, Nadeem, Zelcer, John, and Wickramasinghe, Nilmini
- Abstract
In the context of dementia care, Artificial Intelligence (AI) powered clinical decision support systems have the potential to enhance diagnosis and management. However, the scope and challenges of applying these technologies remain unclear. This scoping review aims to investigate the current state of AI applications in the development of intelligent decision support systems for dementia care. We conducted a comprehensive scoping review of empirical studies that utilised AI-powered clinical decision support systems in dementia care. The results indicate that AI applications in dementia care primarily focus on diagnosis, with limited attention to other aspects outlined in the World Health Organization (WHO) Global Action Plan on the Public Health Response to Dementia 2017–2025 (GAPD). A trifecta of challenges, encompassing data availability, cost considerations, and AI algorithm performance, emerges as noteworthy barriers in adoption of AI applications in dementia care. To address these challenges and enhance AI reliability, we propose a novel approach: a digital twin-based patient journey model. Future research should address identified gaps in GAPD action areas, navigate data-related obstacles, and explore the implementation of digital twins. Additionally, it is imperative to emphasize that addressing trust and combating the stigma associated with AI in healthcare should be a central focus of future research directions. • This paper examines the trajectory of AI-based CDSS application in dementia care. • It focuses on GAPD's seven action areas, highlighting AI's role in improving diagnosis and care. • It identifies limitations in data volume and cost as barriers to their development and usage in practice. • A model utilising digital twin technology is presented as a potential approach to addressing these identified limitations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
20. GCNGAT: Drug–disease association prediction based on graph convolution neural network and graph attention network.
- Author
-
Yang, Runtao, Fu, Yao, Zhang, Qian, and Zhang, Lina
- Abstract
Predicting drug–disease associations can contribute to discovering new therapeutic potentials of drugs, and providing important association information for new drug research and development. Many existing drug–disease association prediction methods have not distinguished relevant background information for the same drug targeted to different diseases. Therefore, this paper proposes a drug–disease association prediction model based on graph convolutional network and graph attention network (GCNGAT) to reposition marketed drugs under the distinguishment of background information. Firstly, in order to obtain initial drug–disease information, a drug–disease heterogeneous graph structure is constructed based on all known drug–disease associations. Secondly, based on the heterogeneous graph structure, the corresponding subgraphs of each group of drug–disease association pairs are extracted to distinguish different background information for the same drug from different diseases. Finally, a model combining Graph neural network with global Average pooling (GnnAp) is designed to predict potential drug–disease associations by learning drug–disease interaction feature representations. The experimental results show that adding subgraph extraction can effectively improve the prediction performance of the model, and the graph representation learning module can fully extract the deep features of drug–disease. Using the 5-fold cross-validation, the proposed model (GCNGAT) achieves AUC (Area Under the receiver operating characteristic Curve) values of 0.9182 and 0.9417 on the PREDICT dataset and CDataset dataset, respectively. Compared with other predictors on the same dataset (PREDICT dataset), GCNGAT outperforms the existing best-performing model (PSGCN), with a 1.58% increase in the AUC value. It is anticipated that this model can provide experimental reference for drug repositioning and further promote the drug research and development process. • A drug–disease heterogeneous graph structure is constructed. • Subgraphs of association pairs are extracted to distinguish background information. • A model combining a graph convolution network and a graph attention network is built. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. GravityNet for end-to-end small lesion detection.
- Author
-
Russo, Ciro, Bria, Alessandro, and Marrocco, Claudio
- Abstract
This paper introduces a novel one-stage end-to-end detector specifically designed to detect small lesions in medical images. Precise localization of small lesions presents challenges due to their appearance and the diverse contextual backgrounds in which they are found. To address this, our approach introduces a new type of pixel-based anchor that dynamically moves towards the targeted lesion for detection. We refer to this new architecture as GravityNet, and the novel anchors as gravity points since they appear to be "attracted" by the lesions. We conducted experiments on two well-established medical problems involving small lesions to evaluate the performance of the proposed approach: microcalcifications detection in digital mammograms and microaneurysms detection in digital fundus images. Our method demonstrates promising results in effectively detecting small lesions in these medical imaging tasks. • Localization of small lesions is a challenge in several medical applications. • GravityNet is a one-stage end-to-end detector for small lesion detection. • Gravity points are a new type of pixel-based anchoring technique. • Gravity points are "attracted" by the targeted lesion for detection. • Experiments demonstrate promising results in detecting small lesions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. A label information fused medical image report generation framework.
- Author
-
Sun, Shuifa, Mei, Zhoujunsen, Li, Xiaolong, Tang, Tinglong, Su, Zhanglin, and Wu, Yirong
- Abstract
Medical imaging is an important tool for clinical diagnosis. Nevertheless, it is very time-consuming and error-prone for physicians to prepare imaging diagnosis reports. Therefore, it is necessary to develop some methods to generate medical imaging reports automatically. Currently, the task of medical imaging report generation is challenging in at least two aspects: (1) medical images are very similar to each other. The differences between normal and abnormal images and between different abnormal images are usually trivial; (2) unrelated or incorrect keywords describing abnormal findings in the generated reports lead to mis-communications. In this paper, we propose a medical image report generation framework composed of four modules, including a Transformer encoder, a MIX-MLP multi-label classification network, a co-attention mechanism (CAM) based semantic and visual feature fusion, and a hierarchical LSTM decoder. The Transformer encoder can be used to learn long-range dependencies between images and labels, effectively extract visual and semantic features of images, and establish long-term dependent relationships between visual and semantic information to accurately extract abnormal features from images. The MIX-MLP multi-label classification network, the co-attention mechanism and the hierarchical LSTM network can better identify abnormalities, achieving visual and text alignment fusion and multi-label diagnostic classification to better facilitate report generation. The results of the experiments performed on two widely used radiology report datasets, IU X-RAY and MIMIC-CXR, show that our proposed framework outperforms current report generation models in terms of both natural linguistic generation metrics and clinical efficacy assessment metrics. The code of this work is available online at https://github.com/watersunhznu/LIFMRG. • A label information fused medical image report generation framework. • ViT encoder based low-level feature extraction for medical image report generation. • A MIX-MLP based multi-label classification network for medical image tag classification. • A POS-SCAN based co-attention mechanism that aligns and fuses both visual information and semantic information. • The focal loss function for the multi-label classification network to address the problem of imbalanced and sparse label distribution. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. Never tell me the odds: Investigating pro-hoc explanations in medical decision making.
- Author
-
Cabitza, Federico, Natali, Chiara, Famiglini, Lorenzo, Campagner, Andrea, Caccavella, Valerio, and Gallazzi, Enrico
- Abstract
This paper examines a kind of explainable AI, centered around what we term pro-hoc explanations , that is a form of support that consists of offering alternative explanations (one for each possible outcome) instead of a specific post-hoc explanation following specific advice. Specifically, our support mechanism utilizes explanations by examples , featuring analogous cases for each category in a binary setting. Pro-hoc explanations are an instance of what we called frictional AI , a general class of decision support aimed at achieving a useful compromise between the increase of decision effectiveness and the mitigation of cognitive risks, such as over-reliance, automation bias and deskilling. To illustrate an instance of frictional AI, we conducted an empirical user study to investigate its impact on the task of radiological detection of vertebral fractures in x-rays. Our study engaged 16 orthopedists in a 'human-first, second-opinion' interaction protocol. In this protocol, clinicians first made initial assessments of the x-rays without AI assistance and then provided their final diagnosis after considering the pro-hoc explanations. Our findings indicate that physicians, particularly those with less experience, perceived pro-hoc XAI support as significantly beneficial, even though it did not notably enhance their diagnostic accuracy. However, their increased confidence in final diagnoses suggests a positive overall impact. Given the promisingly high effect size observed, our results advocate for further research into pro-hoc explanations specifically, and into the broader concept of frictional AI. • Pro-hoc AI explanations aim to reduce over-reliance, not just give advice. • Study: 16 orthopedists assess pro-hoc explanations in fracture detection. • Introducing Frictional AI: Systems embedding cognitive friction for better decisions. • Pro-hoc XAI aids doctors, enhancing decision-making for novices without accuracy loss. • Research needed on pro-hoc explanations, Frictional AI for decision-making balance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. Patient-specific game-based transfer method for Parkinson's disease severity prediction.
- Author
-
Xue, Zaifa, Lu, Huibin, Zhang, Tao, and Little, Max A.
- Abstract
Dysphonia is one of the early symptoms of Parkinson's disease (PD). Most existing methods use feature selection methods to find the optimal subset of voice features for all PD patients. Few have considered the heterogeneity between patients, which implies the need to provide specific prediction models for different patients. However, building the specific model faces the challenge of small sample size, which makes it lack generalization ability. Instance transfer is an effective way to solve this problem. Therefore, this paper proposes a patient-specific game-based transfer (PSGT) method for PD severity prediction. First, a selection mechanism is used to select PD patients with similar disease trends to the target patient from the source domain, which reduces the risk of negative transfer. Then, the contribution of the transferred subjects and their instances to the disease estimation of the target subject is fairly evaluated by the Shapley value, which improves the interpretability of the method. Next, the proportion of valid instances in the transferred subjects is determined, and the instances with higher contribution are transferred to further reduce the difference between the transferred instance subset and the target subject. Finally, the selected subset of instances is added to the training set of the target subject, and the extended data is fed into the random forest to improve the performance of the method. Parkinson's telemonitoring dataset is used to evaluate the feasibility and effectiveness. The mean values of mean absolute error, root mean square error, and volatility obtained by predicting motor-UPDRS and total-UPDRS for target patients are 1.59, 1.95, 1.56 and 1.98, 2.54, 1.94, respectively. Experiment results show that the PSGT has better performance in both prediction error and stability over compared methods. • A novel model is proposed to predict the severity of Parkinson's disease. • Subjects similar to target patients are selected by subject transfer mechanism. • Shapley value evaluates the contribution of transferred subjects and their instances. • We achieve better prediction performance compared to the state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. Online biomedical named entities recognition by data and knowledge-driven model.
- Author
-
Cao, Lulu, Wu, Chaochen, Luo, Guan, Guo, Chao, and Zheng, Anni
- Abstract
Named entity recognition (NER) is an important task for the natural language processing of biomedical text. Currently, most NER studies standardized biomedical text, but NER for unstandardized biomedical text draws less attention from researchers. Named entities in online biomedical text exist with errors and polymorphisms, which negatively impact NER models' performance and impede support from knowledge representation methods. In this paper, we propose a neural network method that can effectively recognize entities in unstandardized online medical/health text. We introduce a new pre-training scheme that uses large-scale online question-answering pairs to enhance transformers' model capacity on online biomedical text. Moreover, we supply models with knowledge representations from a knowledge base called multi-channel knowledge labels, and this method overcomes the restriction from languages, like Chinese, that require word segmentation tools to represent knowledge. Our model outperforms other baseline methods significantly in experiments on a dataset for Chinese online medical entity recognition and achieves state-of-the-art results. [Display omitted] [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
26. Clinical knowledge-guided deep reinforcement learning for sepsis antibiotic dosing recommendations.
- Author
-
Wang, Yuan, Liu, Anqi, Yang, Jucheng, Wang, Lin, Xiong, Ning, Cheng, Yisong, and Wu, Qin
- Abstract
Sepsis is the third leading cause of death worldwide. Antibiotics are an important component in the treatment of sepsis. The use of antibiotics is currently facing the challenge of increasing antibiotic resistance (Evans et al., 2021). Sepsis medication prediction can be modeled as a Markov decision process, but existing methods fail to integrate with medical knowledge, making the decision process potentially deviate from medical common sense and leading to underperformance. (Wang et al., 2021). In this paper, we use Deep Q-Network (DQN) to construct a Sepsis Anti-infection DQN (SAI-DQN) model to address the challenge of determining the optimal combination and duration of antibiotics in sepsis treatment. By setting sepsis clinical knowledge as reward functions to guide DQN complying with medical guidelines, we formed personalized treatment recommendations for antibiotic combinations. The results showed that our model had a higher average value for decision-making than clinical decisions. For the test set of patients, our model predicts that 79.07% of patients will achieve a favorable prognosis with the recommended combination of antibiotics. By statistically analyzing decision trajectories and drug action selection, our model was able to provide reasonable medication recommendations that comply with clinical practices. Our model was able to improve patient outcomes by recommending appropriate antibiotic combinations in line with certain clinical knowledge. [Display omitted] • Using deep reinforcement learning to provide antibiotic treatment recommendations for sepsis. • Integrated with clinical data to formulate highly targeted antibiotic combination recommendations. • Guided by clinical expert knowledge and relevant medical guidelines. • Improving patient prognosis and reducing duration of antibiotic use. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. CSCA U-Net: A channel and space compound attention CNN for medical image segmentation.
- Author
-
Shu, Xin, Wang, Jiashu, Zhang, Aoping, Shi, Jinlong, and Wu, Xiao-Jun
- Abstract
Image segmentation is one of the vital steps in medical image analysis. A large number of methods based on convolutional neural networks have emerged, which can extract abstract features from multiple-modality medical images, learn valuable information that is difficult to recognize by humans, and obtain more reliable results than traditional image segmentation approaches. U-Net, due to its simple structure and excellent performance, is widely used in medical image segmentation. In this paper, to further improve the performance of U-Net, we propose a channel and space compound attention (CSCA) convolutional neural network, CSCA U-Net in abbreviation, which increases the network depth and employs a double squeeze-and-excitation (DSE) block in the bottleneck layer to enhance feature extraction and obtain more high-level semantic features. Moreover, the characteristics of the proposed method are three-fold: (1) channel and space compound attention (CSCA) block, (2) cross-layer feature fusion (CLFF), and (3) deep supervision (DS). Extensive experiments on several available medical image datasets, including Kvasir-SEG, CVC-ClinicDB, CVC-ColonDB, ETIS, CVC-T, 2018 Data Science Bowl (2018 DSB), ISIC 2018, and JSUAH-Cerebellum, show that CSCA U-Net achieves competitive results and significantly improves generalization performance. The codes and trained models are available at https://github.com/xiaolanshu/CSCA-U-Net. • CSCA U-Net is a six-layer U-shaped network that enhances expressiveness. • CSCA U-Net achieves a greater receptive field at the bottleneck layer. • DSE enhances feature extraction ability and obtains more advanced semantic features. • CSCA block makes the model focus on target regions in the feature map. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Detecting mental and physical disorders using multi-task learning equipped with knowledge graph attention network.
- Author
-
Zhang, Wei, Kong, Ling, Lee, Soobin, Chen, Yan, Zhang, Guangxu, Wang, Hao, and Song, Min
- Abstract
Mental and physical disorders (MPD) are inextricably linked in many medical cases; psychosomatic diseases can be induced by mental concerns and psychological discomfort can ensue from physiological diseases. However, existing medical informatics studies focus on identifying mental or physical disorders from a unilateral perspective. Consequently, no existing domain knowledge base, corpus, or detection modeling approach considers mental as well as physical aspects concurrently. This paper proposes a joint modeling approach to detect MPD. First, we crawl through online medical consultation records of patients from websites and build an MPD knowledge ontology by extracting the core conceptual features of the text. Based on the ontology, an MPD knowledge graph containing 12,673 nodes and 82,195 relations is obtained using term matching with a domain thesaurus of each concept. Subsequently, an MPD corpus with fine-grained severities (None , Mild , Moderate , Severe , Dangerous) and 8909 records is constructed by formulating MPD classification criteria and a data annotation process under the guidance of domain experts. Taking the knowledge graph and corpus as the dataset, we design a multi-task learning model to detect the MPD severity, in which a knowledge graph attention network (KGAT) is embedded to better extract knowledge features. Experiments are performed to demonstrate the effectiveness of our model. Furthermore, we employ ontology-based and centrality-based methods to discover additional potential inferred knowledge, which can be captured by KGAT so as to improve the prediction performance and interpretability of our model. Our dataset has been made publicly available, so it can be further used as a medical informatics reference in the fields of psychosomatic medicine, psychiatrics, physical co-morbidity, and so on. • We construct the first corpus combining the fine-grained severity of mental disorders and physical disorders. • We are the first to construct a large-scale mental and physical disorder knowledge graph. • We present a multi-task learning model to automatically detect the mental and physical disorder severity of patients. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. DISCOVER: 2-D multiview summarization of Optical Coherence Tomography Angiography for automatic diabetic retinopathy diagnosis.
- Author
-
El Habib Daho, Mostafa, Li, Yihao, Zeghlache, Rachid, Boité, Hugo Le, Deman, Pierre, Borderie, Laurent, Ren, Hugang, Mannivanan, Niranchana, Lepicard, Capucine, Cochener, Béatrice, Couturier, Aude, Tadayoni, Ramin, Conze, Pierre-Henri, Lamard, Mathieu, and Quellec, Gwenolé
- Abstract
Diabetic Retinopathy (DR), an ocular complication of diabetes, is a leading cause of blindness worldwide. Traditionally, DR is monitored using Color Fundus Photography (CFP), a widespread 2-D imaging modality. However, DR classifications based on CFP have poor predictive power, resulting in suboptimal DR management. Optical Coherence Tomography Angiography (OCTA) is a recent 3-D imaging modality offering enhanced structural and functional information (blood flow) with a wider field of view. This paper investigates automatic DR severity assessment using 3-D OCTA. A straightforward solution to this task is a 3-D neural network classifier. However, 3-D architectures have numerous parameters and typically require many training samples. A lighter solution consists in using 2-D neural network classifiers processing 2-D en-face (or frontal) projections and/or 2-D cross-sectional slices. Such an approach mimics the way ophthalmologists analyze OCTA acquisitions: (1) en-face flow maps are often used to detect avascular zones and neovascularization, and (2) cross-sectional slices are commonly analyzed to detect macular edemas, for instance. However, arbitrary data reduction or selection might result in information loss. Two complementary strategies are thus proposed to optimally summarize OCTA volumes with 2-D images: (1) a parametric en-face projection optimized through deep learning and (2) a cross-sectional slice selection process controlled through gradient-based attribution. The full summarization and DR classification pipeline is trained from end to end. The automatic 2-D summary can be displayed in a viewer or printed in a report to support the decision. We show that the proposed 2-D summarization and classification pipeline outperforms direct 3-D classification with the advantage of improved interpretability. • An algorithm to quickly and efficiently classify 3-D images is presented. • 3-D images are summarized by one or a few 2-D images as an intermediate step. • The pipeline is trained from end to end: the 2-D summary is thus problem-specific. • This multiview 2-D summarization allows interpretability and knowledge discovery. • It is applied to diabetic retinopathy severity assessment using OCT angiography. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Triplet-branch network with contrastive prior-knowledge embedding for disease grading.
- Author
-
Li, Yuexiang, Wang, Yanping, Lin, Guang, Huang, Yawen, Liu, Jingxin, Lin, Yi, Wei, Dong, Zhang, Qirui, Ma, Kai, Zhang, Zhiqiang, Lu, Guangming, and Zheng, Yefeng
- Abstract
Since different disease grades require different treatments from physicians, i.e., the low-grade patients may recover with follow-up observations whereas the high-grade may need immediate surgery, the accuracy of disease grading is pivotal in clinical practice. In this paper, we propose a Triplet-Branch Network with ContRastive priOr-knoWledge embeddiNg (TBN-CROWN) for the accurate disease grading, which enables physicians to accordingly take appropriate treatments. Specifically, our TBN-CROWN has three branches, which are implemented for representation learning, classifier learning and grade-related prior-knowledge learning, respectively. The former two branches deal with the issue of class-imbalanced training samples, while the latter one embeds the grade-related prior-knowledge via a novel auxiliary module, termed contrastive embedding module. The proposed auxiliary module takes the features embedded by different branches as input, and accordingly constructs positive and negative embeddings for the model to deploy grade-related prior-knowledge via contrastive learning. Extensive experiments on our private and two publicly available disease grading datasets show that our TBN-CROWN can effectively tackle the class-imbalance problem and yield a satisfactory grading accuracy for various diseases, such as fatigue fracture, ulcerative colitis, and diabetic retinopathy. [Display omitted] • The auxiliary ranking task in previous TBN is improved to a novel contrastive embedding module. • The proposed TBN-CROWN is validated on the publicly available ulcerative colitis grading dataset. • This is the first work trying to develop an automated system for fatigue fracture diagnosis. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. Scalable Swin Transformer network for brain tumor segmentation from incomplete MRI modalities.
- Author
-
Zhang, Dongsong, Wang, Changjian, Chen, Tianhua, Chen, Weidao, and Shen, Yiqing
- Abstract
Deep learning methods have shown great potential in processing multi-modal Magnetic Resonance Imaging (MRI) data, enabling improved accuracy in brain tumor segmentation. However, the performance of these methods can suffer when dealing with incomplete modalities, which is a common issue in clinical practice. Existing solutions, such as missing modality synthesis, knowledge distillation, and architecture-based methods, suffer from drawbacks such as long training times, high model complexity, and poor scalability. This paper proposes IMS2Trans, a novel lightweight scalable Swin Transformer network by utilizing a single encoder to extract latent feature maps from all available modalities. This unified feature extraction process enables efficient information sharing and fusion among the modalities, resulting in efficiency without compromising segmentation performance even in the presence of missing modalities. Two datasets, BraTS 2018 and BraTS 2020, containing incomplete modalities for brain tumor segmentation are evaluated against popular benchmarks. On the BraTS 2018 dataset, our model achieved higher average Dice similarity coefficient (DSC) scores for the whole tumor, tumor core, and enhancing tumor regions (86.57, 75.67, and 58.28, respectively), in comparison with a state-of-the-art model, i.e. mmFormer (86.45, 75.51, and 57.79, respectively). Similarly, on the BraTS 2020 dataset, our model scored higher DSC scores in these three brain tumor regions (87.33, 79.09, and 62.11, respectively) compared to mmFormer (86.17, 78.34, and 60.36, respectively). We also conducted a Wilcoxon test on the experimental results, and the generated p -value confirmed that our model's performance was statistically significant. Moreover, our model exhibits significantly reduced complexity with only 4.47 M parameters, 121.89G FLOPs, and a model size of 77.13 MB, whereas mmFormer comprises 34.96 M parameters, 265.79 G FLOPs, and a model size of 559.74 MB. These indicate our model, being light-weighted with significantly reduced parameters, is still able to achieve better performance than a state-of-the-art model. By leveraging a single encoder for processing the available modalities, IMS2Trans offers notable scalability advantages over methods that rely on multiple encoders. This streamlined approach eliminates the need for maintaining separate encoders for each modality, resulting in a lightweight and scalable network architecture. The source code of IMS2Trans and the associated weights are both publicly available at https://github.com/hudscomdz/IMS2Trans. • We propose a scalable Swin Transformer for the incomplete MRI modalities brain tumor segmentation. • We introduce a lightweight MLP bottleneck that not only reduces model parameters but also obtains better feature maps. • We introduce a feature distillation to improve consistency across different modalities. • We propose a 3D multimodal CutMix specifically for multi-modal MRI to enhance the robustness. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
32. Opportunities and challenges of artificial intelligence and distributed systems to improve the quality of healthcare service.
- Author
-
Aminizadeh, Sarina, Heidari, Arash, Dehghan, Mahshid, Toumaj, Shiva, Rezaei, Mahsa, Jafari Navimipour, Nima, Stroppa, Fabio, and Unal, Mehmet
- Abstract
The healthcare sector, characterized by vast datasets and many diseases, is pivotal in shaping community health and overall quality of life. Traditional healthcare methods, often characterized by limitations in disease prevention, predominantly react to illnesses after their onset rather than proactively averting them. The advent of Artificial Intelligence (AI) has ushered in a wave of transformative applications designed to enhance healthcare services, with Machine Learning (ML) as a noteworthy subset of AI. ML empowers computers to analyze extensive datasets, while Deep Learning (DL), a specific ML methodology, excels at extracting meaningful patterns from these data troves. Despite notable technological advancements in recent years, the full potential of these applications within medical contexts remains largely untapped, primarily due to the medical community's cautious stance toward novel technologies. The motivation of this paper lies in recognizing the pivotal role of the healthcare sector in community well-being and the necessity for a shift toward proactive healthcare approaches. To our knowledge, there is a notable absence of a comprehensive published review that delves into ML, DL and distributed systems, all aimed at elevating the Quality of Service (QoS) in healthcare. This study seeks to bridge this gap by presenting a systematic and organized review of prevailing ML, DL, and distributed system algorithms as applied in healthcare settings. Within our work, we outline key challenges that both current and future developers may encounter, with a particular focus on aspects such as approach, data utilization, strategy, and development processes. Our study findings reveal that the Internet of Things (IoT) stands out as the most frequently utilized platform (44.3 %), with disease diagnosis emerging as the predominant healthcare application (47.8 %). Notably, discussions center significantly on the prevention and identification of cardiovascular diseases (29.2 %). The studies under examination employ a diverse range of ML and DL methods, along with distributed systems, with Convolutional Neural Networks (CNNs) being the most commonly used (16.7 %), followed by Long Short-Term Memory (LSTM) networks (14.6 %) and shallow learning networks (12.5 %). In evaluating QoS, the predominant emphasis revolves around the accuracy parameter (80 %). This study highlights how ML, DL, and distributed systems reshape healthcare. It contributes to advancing healthcare quality, bridging the gap between technology and medical adoption, and benefiting practitioners and patients. • Conducting a systematic review of available technologies for enhancing healthcare QoS; • Performing a thorough analysis of current challenges related to DL methods in the healthcare sector; • Exploring novel applications of AI in healthcare; • Investigating the impact of regulatory changes on healthcare technology adoption; • Evaluating each platform of distributed systems by highlighting various attributes, including advantages, challenges, datasets and dataset size, usages, privacy, and security; • Describing vital aspects where the preceding strategies may be improved shortly;•Investigating emerging trends in healthcare technology; • Assessing the scalability of distributed systems in healthcare applications. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
33. Assessment and treatment of visuospatial neglect using active learning with Gaussian processes regression.
- Author
-
De Boi, Ivan, Embrechts, Elissa, Schatteman, Quirine, Penne, Rudi, Truijen, Steven, and Saeys, Wim
- Abstract
Visuospatial neglect is a disorder characterised by impaired awareness for visual stimuli located in regions of space and frames of reference. It is often associated with stroke. Patients can struggle with all aspects of daily living and community participation. Assessment methods are limited and show several shortcomings, considering they are mainly performed on paper and do not implement the complexity of daily life. Similarly, treatment options are sparse and often show only small improvements. We present an artificial intelligence solution designed to accurately assess a patient's visuospatial neglect in a three-dimensional setting. We implement an active learning method based on Gaussian process regression to reduce the effort it takes a patient to undergo an assessment. Furthermore, we describe how this model can be utilised in patient oriented treatment and how this opens the way to gamification, tele-rehabilitation and personalised healthcare, providing a promising avenue for improving patient engagement and rehabilitation outcomes. To validate our assessment module, we conducted clinical trials involving patients in a real-world setting. We compared the results obtained using our AI-based assessment with the widely used conventional visuospatial neglect tests currently employed in clinical practice. The validation process serves to establish the accuracy and reliability of our model, confirming its potential as a valuable tool for diagnosing and monitoring visuospatial neglect. Our VR application proves to be more sensitive, while intra-rater reliability remains high. [Display omitted] • We formulate an artificial intelligence framework designed to assess search behaviour and attentional deficits in a virtual 3D setting. • We implement an active learning method based on Gaussian process regression to reduce the effort it takes a patient to undergo this assessment. • We describe how this model can be utilised in treatment and how this opens the way to gamification, tele-rehabilitation and personalised healthcare. • We describe our VR application which demonstrates these practises. • We validate our findings on patients in a clinical setting. We compare them to the standard tests used in practice today. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
34. Multi input–Multi output 3D CNN for dementia severity assessment with incomplete multimodal data.
- Author
-
Gravina, Michela, García-Pedrero, Angel, Gonzalo-Martín, Consuelo, Sansone, Carlo, and Soda, Paolo
- Abstract
Alzheimer's Disease is the most common cause of dementia, whose progression spans in different stages, from very mild cognitive impairment to mild and severe conditions. In clinical trials, Magnetic Resonance Imaging (MRI) and Positron Emission Tomography (PET) are mostly used for the early diagnosis of neurodegenerative disorders since they provide volumetric and metabolic function information of the brain, respectively. In recent years, Deep Learning (DL) has been employed in medical imaging with promising results. Moreover, the use of the deep neural networks, especially Convolutional Neural Networks (CNNs), has also enabled the development of DL-based solutions in domains characterized by the need of leveraging information coming from multiple data sources, raising the Multimodal Deep Learning (MDL). In this paper, we conduct a systematic analysis of MDL approaches for dementia severity assessment exploiting MRI and PET scans. We propose a Multi Input–Multi Output 3D CNN whose training iterations change according to the characteristic of the input as it is able to handle incomplete acquisitions, in which one image modality is missed. Experiments performed on OASIS-3 dataset show the satisfactory results of the implemented network, which outperforms approaches exploiting both single image modality and different MDL fusion techniques. [Display omitted] • Evaluation of multimodal deep learning approaches for dementia severity assessment. • Training strategy to manage incomplete dataset in multimodal deep learning. • Multi input-multi output 3D CNN to process brain MRI and PET acquisitions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Finding neural correlates of depersonalisation/derealisation disorder via explainable CNN-based analysis guided by clinical assessment scores.
- Author
-
Salami, Abbas, Andreu-Perez, Javier, and Gillmeister, Helge
- Abstract
Mental health disorders are typically diagnosed based on subjective reports (e.g., through questionnaires) followed by clinical interviews to evaluate the self-reported symptoms. Therefore, considering the interconnected nature of psychiatric disorders, their accurate diagnosis is a real challenge without indicators of underlying physiological dysfunction. Depersonalisation/derealisation disorder (DPD) is an example of dissociative disorder affecting 1–2 % of the population. DPD is characterised mainly by persistent disembodiment, detachment from surroundings, and feelings of emotional numbness, which can significantly impact patients' quality of life. The underlying neural correlates of DPD have been investigated for years to understand and help with a more accurate and in-time diagnosis of the disorder. However, in terms of EEG studies, which hold great importance due to their convenient and inexpensive nature, the literature has often been based on hypotheses proposed by experts in the field, which require prior knowledge of the disorder. In addition, participants' labelling in research experiments is often derived from the outcome of the Cambridge Depersonalisation Scale (CDS), a subjective assessment to quantify the level of depersonalisation/derealisation, the threshold and reliability of which might be challenged. As a result, we aimed to propose a novel end-to-end EEG processing pipeline based on deep neural networks for DPD biomarker discovery, which requires no prior handcrafted labelled data. Alternatively, it can assimilate knowledge from clinical outcomes like CDS as well as data-driven patterns that differentiate individual brain responses. In addition, the structure of the proposed model targets the uncertainty in CDS scores by using them as prior information only to guide the unsupervised learning task in a multi-task learning scenario. A comprehensive evaluation has been done to confirm the significance of the proposed deep structure, including new ways of network visualisation to investigate spectral, spatial, and temporal information derived in the learning process. We argued that the proposed EEG analytics could also be applied to investigate other psychological and mental disorders currently indicated on the basis of clinical assessment scores. The code to reproduce the results presented in this paper is openly accessible at https://github.com/AbbasSalami/DPD_Analysis. [Display omitted] • The first EEG-based biomarker discovery system for depersonalisation/derealisation disorder is proposed. • The system is made up of a novel multi-input multi-output deep CNN, which requires no prior knowledge of the disorder. • An explainable visualisation approach is proposed to investigate the neural correlates associated with DPD symptoms. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. Overlapping cytoplasms segmentation via constrained multi-shape evolution for cervical cancer screening.
- Author
-
Song, Youyi, Zhang, Ao, Zhou, Jinglin, Luo, Yu, Lin, Zhizhe, and Zhou, Teng
- Abstract
Segmenting overlapping cytoplasms in cervical smear images is a clinically essential task for quantitatively measuring cell-level features to screen cervical cancer This task, however, remains rather challenging, mainly due to the deficiency of intensity (or color) information in the overlapping region Although shape prior-based models that compensate intensity deficiency by introducing prior shape information about cytoplasm are firmly established, they often yield visually implausible results, as they model shape priors only by limited shape hypotheses about cytoplasm, exploit cytoplasm-level shape priors alone, and impose no shape constraint on the resulting shape of the cytoplasm In this paper, we present an effective shape prior-based approach, called constrained multi-shape evolution, that segments all overlapping cytoplasms in the clump simultaneously by jointly evolving each cytoplasm's shape guided by the modeled shape priors We model local shape priors (cytoplasm–level) by an infinitely large shape hypothesis set which contains all possible shapes of the cytoplasm In the shape evolution, we compensate intensity deficiency for the segmentation by introducing not only the modeled local shape priors but also global shape priors (clump–level) modeled by considering mutual shape constraints of cytoplasms in the clump We also constrain the resulting shape in each evolution to be in the built shape hypothesis set for further reducing implausible segmentation results We evaluated the proposed method in two typical cervical smear datasets, and the extensive experimental results confirm its effectiveness. [Display omitted] • A new algorithm to segment overlapping cervical cytoplasms. • An effective scheme to exploit shape priors to compensate intensity information. • A learning framework to automatically capture shape priors. • The considerable performance improvement. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Prediction on nature of cancer by fuzzy graphoidal covering number using artificial neural network.
- Author
-
Bhattacharya, Anushree and Pal, Madhumangal
- Abstract
Predicting the chances of various types of cancers for different organs in the human body is a typical decision-making process in medicine and health. The signaling pathways have played a vital role in increasing or decreasing the possibility of the deadliest disease, cancer. To combine the pathways concept and ambiguity in the prediction techniques of such diseases, we have used the proposed research on fuzzy graphoidal covers of fuzzy graphs in this paper. Determining a path with uncertainty and shortest length is a challenging topic of graph theory, and a collection of such shortest paths maintaining specific conditions is defined as a fuzzy graphoidal cover for a fuzzy graph. Also, we have defined fuzzy graphoidal covering number as a new parameter, reflecting the measure of coverage by fuzzy graphoidal covering set in a system. Afterwards, some important characterizations of the fuzzy graphoidal covering number are established with justified proof. Also, specific limit values of this number are provided for particular cases. Then, we developed an efficient algorithm for finding the defined covering set with its space and time complexity. The findings of this proposed study have been composed with an artificial neural network to model a strong tool for resolving an essential issue of medical sciences, the prediction of cancer type in the human body. We have analyzed two types of neural networks such as one one-dimensional and two-dimensional specification, for clarity of the obtained results. Also, we have found out the most possible cancer type is breast cancer from the data of our considered case study as a concluding statement for any decision-maker in the field of health sciences. Finally, sensitivity analysis and comparative study have been done to show the stability of our proposed work. • Introducing new measure for covering a fuzzy graph, fuzzy graphoidal covering number. • Many propositions of fuzzy graphoidal covering numbers are given with proper proofs. • Also, certain bounds are mentioned and proved for specific fuzzy graphs. • Prediction of cancer type in the human body is discussed as a real-life application. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
38. Semi-supervised image segmentation using a residual-driven mean teacher and an exponential Dice loss.
- Author
-
Mei, Chenyang, Yang, Xiaoguo, Zhou, Mi, Zhang, Shaodan, Chen, Hao, Yang, Xiaokai, and Wang, Lei
- Abstract
Semi-supervised segmentation plays an important role in computer vision and medical image analysis and can alleviate the burden of acquiring abundant expert-annotated images. In this paper, we developed a residual-driven semi-supervised segmentation method (termed RDMT) based on the classical mean teacher (MT) framework by introducing a novel model-level residual perturbation and an exponential Dice (eDice) loss. The introduced perturbation was integrated into the exponential moving average (EMA) scheme to enhance the performance of the MT, while the eDice loss was used to improve the detection sensitivity of a given network to object boundaries. We validated the developed method by applying it to segment 3D Left Atrium (LA) and 2D optic cup (OC) from the public LASC and REFUGE datasets based on the V-Net and U-Net, respectively. Extensive experiments demonstrated that the developed method achieved the average Dice score of 0.8776 and 0.7751, when trained on 10% and 20% labeled images, respectively for the LA and OC regions depicted on the LASC and REFUGE datasets. It significantly outperformed the MT and can compete with several existing semi-supervised segmentation methods (i.e., HCMT, UAMT, DTC and SASS). • A residual perturbation and an exponential Dice loss were introduced. • A novel semi-supervised segmentation method was proposed based on the mean teacher framework. • The proposed method was validated on two public datasets via the V-Net and U-Net. • Extensive experiments showed the advantage of the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. Contrastive image adaptation for acquisition shift reduction in medical imaging.
- Author
-
Hognon, Clément, Conze, Pierre-Henri, Bourbonne, Vincent, Gallinato, Olivier, Colin, Thierry, Jaouen, Vincent, and Visvikis, Dimitris
- Abstract
The domain shift, or acquisition shift in medical imaging, is responsible for potentially harmful differences between development and deployment conditions of medical image analysis techniques. There is a growing need in the community for advanced methods that could mitigate this issue better than conventional approaches. In this paper, we consider configurations in which we can expose a learning-based pixel level adaptor to a large variability of unlabeled images during its training, i.e. sufficient to span the acquisition shift expected during the training or testing of a downstream task model. We leverage the ability of convolutional architectures to efficiently learn domain-agnostic features and train a many-to-one unsupervised mapping between a source collection of heterogeneous images from multiple unknown domains subjected to the acquisition shift and a homogeneous subset of this source set of lower cardinality, potentially constituted of a single image. To this end, we propose a new cycle-free image-to-image architecture based on a combination of three loss functions : a contrastive PatchNCE loss, an adversarial loss and an edge preserving loss allowing for rich domain adaptation to the target image even under strong domain imbalance and low data regimes. Experiments support the interest of the proposed contrastive image adaptation approach for the regularization of downstream deep supervised segmentation and cross-modality synthesis models. • Acquisition shift is responsible for lack of robustness of deep models to scanner and protocol changes. • We propose a deep image adaptor to mitigate the acquisition shift, named contrastive image adaptation. • Structural preservation is enforced with a gradient loss while translation is guided by GAN and contrastive loss. • The method requires as little as one 2D image in the target domain. • Experiments show positive impact on segmentation and image to image translation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. Diagnosis knowledge constrained network based on first-order logic for syndrome differentiation.
- Author
-
Li, Meiwen, Wang, Lin, Wu, Qingtao, Zhu, Junlong, and Zhang, Mingchuan
- Abstract
Traditional Chinese medicine (TCM) has been recognized worldwide as a valuable asset of human medicine. The procedure of TCM is to treatment based on syndrome differentiation. However, the effect of TCM syndrome differentiation relies heavily on the experience of doctors. The gratifying progress of machine learning research in recent years has brought new ideas for TCM syndrome differentiation. In this paper, we propose a deep network model for TCM syndrome differentiation, which improves network performance by injecting TCM syndrome differentiation knowledge in the form of first-order logic into the deep network. Experimental results show that the accuracy of our proposed model reaches 89%, which is significantly better than the deep learning model MLP and other traditional machine learning models. In addition, we present the collected and formatted TCM syndrome differentiation (TSD) dataset, which contains more than 40,000 TCM clinical records. Moreover, 45 symptoms ("▪"), 322 patterns("▪"), and more than 500 symptoms are labeled in TSD respectively. To the best of our knowledge, this is the first TCM syndrome differentiation dataset labeling diseases, syndromes and pattern. Such detailed labeling is helpful to explore the relationship between various elements of syndrome differentiation. • Challenges to address the syndrome differentiation issue in TCM. • The TCM diagnostic knowledge is represented by fist-order logic rules. • A TCM dataset includes 40,000 clinical records and 500 labels. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. An in-depth survey on Deep Learning-based Motor Imagery Electroencephalogram (EEG) classification.
- Author
-
Wang, Xianheng, Liesaputra, Veronica, Liu, Zhaobin, Wang, Yi, and Huang, Zhiyi
- Abstract
Electroencephalogram (EEG)-based Brain–Computer Interfaces (BCIs) build a communication path between human brain and external devices. Among EEG-based BCI paradigms, the most commonly used one is motor imagery (MI). As a hot research topic, MI EEG-based BCI has largely contributed to medical fields and smart home industry. However, because of the low signal-to-noise ratio (SNR) and the non-stationary characteristic of EEG data, it is difficult to correctly classify different types of MI-EEG signals. Recently, the advances in Deep Learning (DL) significantly facilitate the development of MI EEG-based BCIs. In this paper, we provide a systematic survey of DL-based MI-EEG classification methods. Specifically, we first comprehensively discuss several important aspects of DL-based MI-EEG classification, covering input formulations, network architectures, public datasets, etc. Then, we summarize problems in model performance comparison and give guidelines to future studies for fair performance comparison. Next, we fairly evaluate the representative DL-based models using source code released by the authors and meticulously analyse the evaluation results. By performing ablation study on the network architecture, we found that (1) effective feature fusion is indispensable for multi-stream CNN-based models. (2) LSTM should be combined with spatial feature extraction techniques to obtain good classification performance. (3) the use of dropout contributes little to improving the model performance, and that (4) adding fully connected layers to the models significantly increases their parameters but it might not improve their performance. Finally, we raise several open issues in MI-EEG classification and provide possible future research directions. • 67 relevant studies are identified and included primarily through a systematic meta-analysis procedure. • A comprehensive review of Deep Learning-based Motor Imagery EEG classification from various perspectives. • 13 typical models are selected, evaluated and analysed to shed light on the development of this field. • Ablation studies investigate how specific design factors impact common network architectures. • Open issues and potential future directions in this field are discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. Predicting stroke outcome: A case for multimodal deep learning methods with tabular and CT Perfusion data.
- Author
-
Borsos, Balázs, Allaart, Corinne G., and van Halteren, Aart
- Abstract
Acute ischemic stroke is one of the leading causes of morbidity and disability worldwide, often followed by a long rehabilitation period. To improve and personalize stroke rehabilitation, it is essential to provide a reliable prognosis to caregivers and patients. Deep learning techniques might improve the predictions by incorporating different data modalities. We present a multimodal approach to predict the functional status of acute ischemic stroke patients after their discharge based on tabular data and CT perfusion imaging. We conducted experiments on tabular, imaging, and multimodal deep learning architectures to predict dichotomized mRS scores 3 months after the event. The dataset was collected from a Dutch hospital and includes 98 CVA patients with a visible occlusion on their CT perfusion scan. Tabular data is based on the Dutch Acute Stroke Audit data, and imaging data consists of summed-up CT perfusion maps. On the tabular data, TabNet outperformed our baselines with an AUC of 0.71, while ResNet-10 on the imaging data performed comparably with an AUC of 0.70. Our implementation of the multimodal DAFT architecture outperforms baselines as well as comparable studies by achieving an 0.75 AUC, and 0.80 F1 score. This was achieved with a final model of less than a hundred thousand optimizable parameters, and a dataset less than half the size of reference papers. Overall, we demonstrate the feasibility of predicting the functional outcome for ischemic stroke patients and the usability of multimodal deep learning architectures for this purpose. • Using multimodal clinical and CT perfusion data improves CVA outcome. • Sequential attention in deep learning on tabular data outperforms baselines. • Implemented data fusion technique with affine transformation of feature maps. • Summing up of CT-perfusion scans outperforms complex preprocessing. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
43. STERN: Attention-driven Spatial Transformer Network for abnormality detection in chest X-ray images.
- Author
-
Rocha, Joana, Pereira, Sofia Cardoso, Pedrosa, João, Campilho, Aurélio, and Mendonça, Ana Maria
- Abstract
Chest X-ray scans are frequently requested to detect the presence of abnormalities, due to their low-cost and non-invasive nature. The interpretation of these images can be automated to prioritize more urgent exams through deep learning models, but the presence of image artifacts, e.g. lettering, often generates a harmful bias in the classifiers and an increase of false positive results. Consequently, healthcare would benefit from a system that selects the thoracic region of interest prior to deciding whether an image is possibly pathologic. The current work tackles this binary classification exercise, in which an image is either normal or abnormal, using an attention-driven and spatially unsupervised S patial T ransform er N etwork (STERN), that takes advantage of a novel domain-specific loss to better frame the region of interest. Unlike the state of the art, in which this type of networks is usually employed for image alignment, this work proposes a spatial transformer module that is used specifically for attention, as an alternative to the standard object detection models that typically precede the classifier to crop out the region of interest. In sum, the proposed end-to-end architecture dynamically scales and aligns the input images to maximize the classifier's performance, by selecting the thorax with translation and non-isotropic scaling transformations, and thus eliminating artifacts. Additionally, this paper provides an extensive and objective analysis of the selected regions of interest, by proposing a set of mathematical evaluation metrics. The results indicate that the STERN achieves similar results to using YOLO-cropped images, with reduced computational cost and without the need for localization labels. More specifically, the system is able to distinguish abnormal frontal images from the CheXpert dataset, with a mean AUC of 85.67% - a 2.55% improvement vs. the 0.98% improvement achieved by the YOLO-based counterpart in comparison to a standard baseline classifier. At the same time, the STERN approach requires less than 2/3 of the training parameters, while increasing the inference time per batch in less than 2 ms. Code available via GitHub. • The proposed spatial transformer allows the system to focus on the thoracic region. • This built-in attention-driven model reduces the negative impact of image artifacts. • A novel loss function and a finetuning stage improve the initial methodology. • A set of proposed metrics evaluate and compare the thoracic region selection. • The end-to-end system outperforms typical object detection followed by classification. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. Robot assisted Fetoscopic Laser Coagulation: Improvements in navigation, re-location and coagulation.
- Author
-
Hernansanz, Albert, Parra, Johanna, Sayols, Narcís, Eixarch, Elisenda, Gratacós, Eduard, and Casals, Alícia
- Abstract
Fetoscopic Laser Coagulation (FLC) for Twin to Twin Transfusion Syndrome is a challenging intervention due to the working conditions: low quality images acquired from a 3 mm fetoscope inside a turbid liquid environment, local view of the placental surface, unstable surgical field and delicate tissue layers. FLC is based on locating, coagulating and reviewing anastomoses over the placenta's surface. The procedure demands the surgeons to generate a mental map of the placenta with the distribution of the anastomoses, maintaining, at the same time, precision in coagulation and protecting the placenta and amniotic sac from potential damages. This paper describes a teleoperated platform with a cognitive-based control that provides assistance to improve patient safety and surgery performance during fetoscope navigation, target re-location and coagulation processes. A comparative study between manual and teleoperated operation, executed in dry laboratory conditions, analyzes basic fetoscopic skills: fetoscope navigation and laser coagulation. Two exercises are proposed: first, fetoscope guidance and precise coagulation. Second, a resolved placenta (all anastomoses are indicated) to evaluate navigation, re-location and coagulation. The results are analyzed in terms of economy of movement, execution time, coagulation accuracy, amount of coagulated placental surface and risk of placenta puncture. In addition, new metrics, based on navigation and coagulation maps evaluate robotic performance. The results validate the developed platform, showing noticeable improvements in all the metrics. [Display omitted] • Robotic platform for TTTS surgery to increase patient safety and surgery outcomes. • Cognitive control based using hierarchical Finite State Machine architecture. • Safe autonomous fetoscope navigation to pre-located Points of Interest. • Assisted fetoscope navigation to prevent placenta puncture and reduce cognitive load. • Navigation Map and placenta's mosaicking to understand vascular structures. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
45. A few-shot disease diagnosis decision making model based on meta-learning for general practice.
- Author
-
Liu, Qianghua, Tian, Yu, Zhou, Tianshu, Lyu, Kewei, Xin, Ran, Shang, Yong, Liu, Ying, Ren, Jingjing, and Li, Jingsong
- Abstract
Diagnostic errors have become the biggest threat to the safety of patients in primary health care. General practitioners, as the "gatekeepers" of primary health care, have a responsibility to accurately diagnose patients. However, many general practitioners have insufficient knowledge and clinical experience in some diseases. Clinical decision making tools need to be developed to effectively improve the diagnostic process in primary health care. The long-tailed class distributions of medical datasets are challenging for many popular decision making models based on deep learning, which have difficulty predicting few-shot diseases. Meta-learning is a new strategy for solving few-shot problems. In this study, a few-shot disease diagnosis decision making model based on a model-agnostic meta-learning algorithm (FSDD-MAML) is proposed. The MAML algorithm is applied in a knowledge graph-based disease diagnosis model to find the optimal model parameters. Moreover, FSDD-MAML can learn learning rates for all modules of the knowledge graph-based disease diagnosis model. For n -way, k -shot learning tasks, the inner loop of FSDD-MAML performs multiple gradient update steps to learn internal features in disease classification tasks using n × k examples, and the outer loop of FSDD-MAML optimizes the meta-objective to find the associated optimal parameters and learning rates. FSDD-MAML is compared with the original knowledge graph-based disease diagnosis model and other meta-learning algorithms based on an abdominal disease dataset. Meta-learning algorithms can greatly improve the performance of models in top-1 evaluation compared with top-3, top-5, and top-10 evaluations. The proposed decision making model FSDD-MAML outperforms all the other models, with a precision@1 of 90.02 %. We achieve state-of-the-art performance in the diagnosis of all diseases, and the prediction performance for few-shot diseases is greatly improved. For the two groups with the fewest examples of diseases, FSDD-MAML achieves relative increases in precision@1 of 29.13 % and 21.63 % compared with the original knowledge graph-based disease diagnosis model. In addition, we analyze the reasoning process of several few-shot disease predictions and provide an explanation for the results. The decision making model based on meta-learning proposed in this paper can support the rapid diagnosis of diseases in general practice and is especially capable of helping general practitioners diagnose few-shot diseases. This study is of profound significance for the exploration and application of meta-learning to few-shot disease assessment in general practice. • A few-shot disease diagnosis modelis proposed for improving the prediction performance in general practice. • The proposed model referring to model-agnostic meta-learning can achieve fast adaption for new diseases using few examples. • Learnable learning rates are set in the inner loop of the model for improving the performance and stability of the model. • The case study demonstrates that the proposed model outperforms the baseline models in few-shot diseases.. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. Prediction of Freezing of Gait in Parkinson's disease based on multi-channel time-series neural network.
- Author
-
Wang B, Hu X, Ge R, Xu C, Zhang J, Gao Z, Zhao S, and Polat K
- Abstract
Freezing of Gait (FOG) is a noticeable symptom of Parkinson's disease, like being stuck in place and increasing the risk of falls. The wearable multi-channel sensor system is an efficient method to predict and monitor the FOG, thus warning the wearer to avoid falls and improving the quality of life. However, the existing approaches for the prediction of FOG mainly focus on a single sensor system and cannot handle the interference between multi-channel wearable sensors. Hence, we propose a novel multi-channel time-series neural network (MCT-Net) approach to merge multi-channel gait features into a comprehensive prediction framework, alerting patients to FOG symptoms in advance. Owing to the causal distributed convolution, MCT-Net is a real-time method available to give optimal prediction earlier and implemented in remote devices. Moreover, intra-channel and inter-channel transformers of MCT-Net extract and integrate different sensor position features into a unified deep learning model. Compared with four other state-of-the-art FOG prediction baselines, the proposed MCT-Net obtains 96.21% in accuracy and 80.46% in F1-score on average 2 s before FOG occurrence, demonstrating the superiority of MCT-Net., Competing Interests: Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (Copyright © 2024 Elsevier B.V. All rights reserved.)
- Published
- 2024
- Full Text
- View/download PDF
47. A systematic literature review on the significance of deep learning and machine learning in predicting Alzheimer's disease.
- Author
-
Kaur A, Mittal M, Bhatti JS, Thareja S, and Singh S
- Abstract
Background: Alzheimer's disease (AD) is the most prevalent cause of dementia, characterized by a steady decline in mental, behavioral, and social abilities and impairs a person's capacity for independent functioning. It is a fatal neurodegenerative disease primarily affecting older adults., Objectives: The purpose of this literature review is to investigate various AD detection techniques, datasets, input modalities, algorithms, libraries, and performance evaluation metrics used to determine which model or strategy may provide superior performance., Method: The initial search yielded 807 papers, but only 100 research articles were chosen after applying the inclusion-exclusion criteria. This SLR analyzed research items published between January 2019 and December 2022. The ACM, Elsevier, IEEE Xplore Digital Library, PubMed, Springer and Taylor & Francis were systematically searched. The current study considers articles that used Magnetic Resonance Imaging (MRI), Positron Emission Tomography (PET), APOe4 genotype, Diffusion Tensor Imaging (DTI) and Cerebrospinal Fluid (CSF) biomarkers. The study was performed following Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) guidelines., Conclusion: According to the literature survey, most studies (n = 76) used the DL strategy. The datasets used by studies were primarily derived from the Alzheimer's Disease Neuroimaging Initiative (ADNI) database. The majority of studies (n = 73) used single-modality neuroimaging data, while the remaining used multi-modal input data. In a multi-modality approach, the combination of MRI and PET scans is commonly preferred. Also, Regarding the algorithm used, Convolution Neural Network (CNN) showed the highest accuracy, 100 %, in classifying AD vs. CN subjects whereas the SVM was the most common ML algorithm, with a maximum accuracy of 99.82 %., Competing Interests: Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (Copyright © 2024 Elsevier B.V. All rights reserved.)
- Published
- 2024
- Full Text
- View/download PDF
48. Enhanced differential evolution algorithm for feature selection in tuberculous pleural effusion clinical characteristics analysis.
- Author
-
Zhou X, Chen Y, Gui W, Heidari AA, Cai Z, Wang M, Chen H, and Li C
- Subjects
- Humans, Adenosine Deaminase metabolism, Leukocyte Count, Pleural Effusion microbiology, Support Vector Machine, Algorithms, Tuberculosis, Pleural diagnosis
- Abstract
Tuberculous pleural effusion poses a significant threat to human health due to its potential for severe disease and mortality. Without timely treatment, it may lead to fatal consequences. Therefore, early identification and prompt treatment are crucial for preventing problems such as chronic lung disease, respiratory failure, and death. This study proposes an enhanced differential evolution algorithm based on colony predation and dispersed foraging strategies. A series of experiments conducted on the IEEE CEC 2017 competition dataset validated the global optimization capability of the method. Additionally, a binary version of the algorithm is introduced to assess the algorithm's ability to address feature selection problems. Comprehensive comparisons of the effectiveness of the proposed algorithm with 8 similar algorithms were conducted using public datasets with feature sizes ranging from 10 to 10,000. Experimental results demonstrate that the proposed method is an effective feature selection approach. Furthermore, a predictive model for tuberculous pleural effusion is established by integrating the proposed algorithm with support vector machines. The performance of the proposed model is validated using clinical records collected from 140 tuberculous pleural effusion patients, totaling 10,780 instances. Experimental results indicate that the proposed model can identify key correlated indicators such as pleural effusion adenosine deaminase, temperature, white blood cell count, and pleural effusion color, aiding in the clinical feature analysis of tuberculous pleural effusion and providing early warning for its treatment and prediction., Competing Interests: Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (Copyright © 2024 Elsevier B.V. All rights reserved.)
- Published
- 2024
- Full Text
- View/download PDF
49. TSOANet: Time-Sensitive Orthogonal Attention Network for medical event prediction.
- Author
-
Chen H, Zhang J, Xiang Y, Lu S, and Tang B
- Subjects
- Humans, Time Factors, Neural Networks, Computer, Electronic Health Records
- Abstract
Medical Event Prediction (MEP) based on Electronic Medical Records (EMR) is an essential and valuable task for healthcare. For a patient, information in the EMR can be organized into a structured sequence, consisting of multiple visits each with details about visit time and various types of medical events. As the time intervals between neighboring visits are irregular and the medical events at different visits can vary significantly, MEP based on EMR is still challenging. Many studies have been proposed to model the irregular time intervals, relations among different types of medical events within each visit and relations among medical events across visits, and reported exciting results. However, most of these studies focus on two out of the three aspects mentioned above, with only a few addressing all the three aspects simultaneously. In this study, we propose a novel network, the Time-Sensitive Orthogonal Attention Network (TSOANet), which can fully utilize the irregular time intervals, relations among different types of medical events within and across visits. In particular, we design two key components: (1) Time-Sensitive Block, used to model the time intervals at both local and global levels to determine the impact of each visit in EMR; (2) Orthogonal Attention Block, used to model relations among different types of medical events within each visit and across visits in two axes, that is, event axis and time axis. Extensive experiments on two public real-world EMR datasets demonstrate that TSOANet outperforms the state-of-the-art models for various prediction tasks, thereby verifying the effectiveness of our approach. The source code of TSOANet is released at https://github.com/chh13502/TSOANet., Competing Interests: Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (Copyright © 2024 Elsevier B.V. All rights reserved.)
- Published
- 2024
- Full Text
- View/download PDF
50. Leveraging VQ-VAE tokenization for autoregressive modeling of medical time series.
- Author
-
Lee Y, Chae Y, and Jung K
- Abstract
In this work, we present CodeAR, a medical time series generative model for electronic health record (EHR) synthesis. CodeAR employs autoregressive modeling on discrete tokens obtained using a vector quantized-variational autoencoder (VQ-VAE), which addresses key challenges of accurate distribution modeling and patient privacy preservation in the medical domain. The proposed model is trained with next-token prediction instead of a regression problem for more accurate distribution modeling, where the autoregressive property of CodeAR is useful to capture the inherent causality in time series data. In addition, the compressive property of the VQ-VAE prevents CodeAR from memorizing the original training data, which ensures patient privacy. Experimental results demonstrate that CodeAR outperforms the baseline autoregressive-based and GAN-based models in terms of maximum mean discrepancy (MMD) and Train on Synthetic, Test on Real tests. Our results highlight the effectiveness of autoregressive modeling on discrete tokens, the utility of CodeAR in causal modeling, and its robustness against data memorization., Competing Interests: Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (Copyright © 2024. Published by Elsevier B.V.)
- Published
- 2024
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.