Back to Search Start Over

A Racially Unbiased, Machine Learning Approach to Prediction of Mortality: Algorithm Development Study

Authors :
Samson Mataraso
Andrea McCoy
Gina Barnes
Ritankar Das
Abigail Green-Saxena
Jacob Calvert
Jana Hoffman
Anna Siefkas
Angier Allen
Emily Pellegrini
Gregory Braden
R. Phillip Dellinger
Hoyt Burdick
Source :
JMIR Public Health and Surveillance, JMIR Public Health and Surveillance, Vol 6, Iss 4, p e22400 (2020)
Publication Year :
2020
Publisher :
JMIR Publications Inc., 2020.

Abstract

Background Racial disparities in health care are well documented in the United States. As machine learning methods become more common in health care settings, it is important to ensure that these methods do not contribute to racial disparities through biased predictions or differential accuracy across racial groups. Objective The goal of the research was to assess a machine learning algorithm intentionally developed to minimize bias in in-hospital mortality predictions between white and nonwhite patient groups. Methods Bias was minimized through preprocessing of algorithm training data. We performed a retrospective analysis of electronic health record data from patients admitted to the intensive care unit (ICU) at a large academic health center between 2001 and 2012, drawing data from the Medical Information Mart for Intensive Careā€“III database. Patients were included if they had at least 10 hours of available measurements after ICU admission, had at least one of every measurement used for model prediction, and had recorded race/ethnicity data. Bias was assessed through the equal opportunity difference. Model performance in terms of bias and accuracy was compared with the Modified Early Warning Score (MEWS), the Simplified Acute Physiology Score II (SAPS II), and the Acute Physiologic Assessment and Chronic Health Evaluation (APACHE). Results The machine learning algorithm was found to be more accurate than all comparators, with a higher sensitivity, specificity, and area under the receiver operating characteristic. The machine learning algorithm was found to be unbiased (equal opportunity difference 0.016, P=.20). APACHE was also found to be unbiased (equal opportunity difference 0.019, P=.11), while SAPS II and MEWS were found to have significant bias (equal opportunity difference 0.038, P=.006 and equal opportunity difference 0.074, P Conclusions This study indicates there may be significant racial bias in commonly used severity scoring systems and that machine learning algorithms may reduce bias while improving on the accuracy of these methods.

Details

ISSN :
23692960
Volume :
6
Database :
OpenAIRE
Journal :
JMIR Public Health and Surveillance
Accession number :
edsair.doi.dedup.....77d7b57ba04de774c34091f8be1cdc72
Full Text :
https://doi.org/10.2196/22400