» Articles » PMID: 39441784

Conceptualizing Bias in EHR Data: A Case Study in Performance Disparities by Demographic Subgroups for a Pediatric Obesity Incidence Classifier

Overview
Date 2024 Oct 23
PMID 39441784
Authors
Affiliations
Soon will be listed here.
Abstract

Electronic Health Records (EHRs) are increasingly used to develop machine learning models in predictive medicine. There has been limited research on utilizing machine learning methods to predict childhood obesity and related disparities in classifier performance among vulnerable patient subpopulations. In this work, classification models are developed to recognize pediatric obesity using temporal condition patterns obtained from patient EHR data in a U.S. study population. We trained four machine learning algorithms (Logistic Regression, Random Forest, Gradient Boosted Trees, and Neural Networks) to classify cases and controls as obesity positive or negative, and optimized hyperparameter settings through a bootstrapping methodology. To assess the classifiers for bias, we studied model performance by population subgroups then used permutation analysis to identify the most predictive features for each model and the demographic characteristics of patients with these features. Mean AUC-ROC values were consistent across classifiers, ranging from 0.72-0.80. Some evidence of bias was identified, although this was through the models performing better for minority subgroups (African Americans and patients enrolled in Medicaid). Permutation analysis revealed that patients from vulnerable population subgroups were over-represented among patients with the most predictive diagnostic patterns. We hypothesize that our models performed better on under-represented groups because the features more strongly associated with obesity were more commonly observed among minority patients. These findings highlight the complex ways that bias may arise in machine learning models and can be incorporated into future research to develop a thorough analytical approach to identify and mitigate bias that may arise from features and within EHR datasets when developing more equitable models.

References
1.
Thakur N, Oh S, Nguyen E, Martin M, Roth L, Galanter J . Socioeconomic status and childhood asthma in urban minority youths. The GALA II and SAGE II studies. Am J Respir Crit Care Med. 2013; 188(10):1202-9. PMC: 3863734. DOI: 10.1164/rccm.201306-1016OC. View

2.
Stern J, Chen M, Fagnano M, Halterman J . Allergic rhinitis co-morbidity on asthma outcomes in city school children. J Asthma. 2022; 60(2):255-261. PMC: 9653514. DOI: 10.1080/02770903.2022.2043363. View

3.
Andaur Navarro C, Damen J, Takada T, Nijman S, Dhiman P, Ma J . Risk of bias in studies on prediction models developed using supervised machine learning techniques: systematic review. BMJ. 2021; 375:n2281. PMC: 8527348. DOI: 10.1136/bmj.n2281. View

4.
Christodoulou E, Ma J, Collins G, Steyerberg E, Verbakel J, Van Calster B . A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J Clin Epidemiol. 2019; 110:12-22. DOI: 10.1016/j.jclinepi.2019.02.004. View

5.
Schuch H, Furtado M, Silva G, Kawachi I, Chiavegatto Filho A, Elani H . Fairness of Machine Learning Algorithms for Predicting Foregone Preventive Dental Care for Adults. JAMA Netw Open. 2023; 6(11):e2341625. PMC: 10625037. DOI: 10.1001/jamanetworkopen.2023.41625. View