Clare C Brown, Michael Thomsen, Benjamin C Amick, J Mick Tilford, Keneshia Bryant-Moore, Horacio Gomez-Acevedo
{"title":"Fairness in Low Birthweight Predictive Models: Implications of Excluding Race/Ethnicity.","authors":"Clare C Brown, Michael Thomsen, Benjamin C Amick, J Mick Tilford, Keneshia Bryant-Moore, Horacio Gomez-Acevedo","doi":"10.1007/s40615-025-02296-x","DOIUrl":null,"url":null,"abstract":"<p><strong>Context: </strong>To evaluate algorithmic fairness in low birthweight predictive models.</p><p><strong>Study design: </strong>This study analyzed insurance claims (n = 9,990,990; 2013-2021) linked with birth certificates (n = 173,035; 2014-2021) from the Arkansas All Payers Claims Database (APCD).</p><p><strong>Methods: </strong>Low birthweight (< 2500 g) predictive models included four approaches (logistic, elastic net, linear discriminate analysis, and gradient boosting machines [GMB]) with and without racial/ethnic information. Model performance was assessed overall, among Hispanic individuals, and among non-Hispanic White, Black, Native Hawaiian/Other Pacific Islander, and Asian individuals using multiple measures of predictive performance (i.e., AUC [area under the receiver operating characteristic curve] scores, calibration, sensitivity, and specificity).</p><p><strong>Results: </strong>AUC scores were lower (underperformed) for Black and Asian individuals relative to White individuals. In the strongest performing model (i.e., GMB), the AUC scores for Black (0.718 [95% CI: 0.705-0.732]) and Asian (0.655 [95% CI: 0.582-0.728]) populations were lower than the AUC for White individuals (0.764 [95% CI: 0.754-0.775 ]). Model performance measured using AUC was comparable in models that included and excluded race/ethnicity; however, sensitivity (i.e., the percent of records correctly predicted as \"low birthweight\" among those who actually had low birthweight) was lower and calibration was weaker, suggesting underprediction for Black individuals when race/ethnicity were excluded.</p><p><strong>Conclusions: </strong>This study found that racially blind models resulted in underprediction and reduced algorithmic performance, measured using sensitivity and calibration, for Black populations. Such under prediction could unfairly decrease resource allocation needed to reduce perinatal health inequities. Population health management programs should carefully consider algorithmic fairness in predictive models and associated resource allocation decisions.</p>","PeriodicalId":16921,"journal":{"name":"Journal of Racial and Ethnic Health Disparities","volume":" ","pages":""},"PeriodicalIF":3.2000,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Racial and Ethnic Health Disparities","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s40615-025-02296-x","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PUBLIC, ENVIRONMENTAL & OCCUPATIONAL HEALTH","Score":null,"Total":0}
引用次数: 0
Abstract
Context: To evaluate algorithmic fairness in low birthweight predictive models.
Study design: This study analyzed insurance claims (n = 9,990,990; 2013-2021) linked with birth certificates (n = 173,035; 2014-2021) from the Arkansas All Payers Claims Database (APCD).
Methods: Low birthweight (< 2500 g) predictive models included four approaches (logistic, elastic net, linear discriminate analysis, and gradient boosting machines [GMB]) with and without racial/ethnic information. Model performance was assessed overall, among Hispanic individuals, and among non-Hispanic White, Black, Native Hawaiian/Other Pacific Islander, and Asian individuals using multiple measures of predictive performance (i.e., AUC [area under the receiver operating characteristic curve] scores, calibration, sensitivity, and specificity).
Results: AUC scores were lower (underperformed) for Black and Asian individuals relative to White individuals. In the strongest performing model (i.e., GMB), the AUC scores for Black (0.718 [95% CI: 0.705-0.732]) and Asian (0.655 [95% CI: 0.582-0.728]) populations were lower than the AUC for White individuals (0.764 [95% CI: 0.754-0.775 ]). Model performance measured using AUC was comparable in models that included and excluded race/ethnicity; however, sensitivity (i.e., the percent of records correctly predicted as "low birthweight" among those who actually had low birthweight) was lower and calibration was weaker, suggesting underprediction for Black individuals when race/ethnicity were excluded.
Conclusions: This study found that racially blind models resulted in underprediction and reduced algorithmic performance, measured using sensitivity and calibration, for Black populations. Such under prediction could unfairly decrease resource allocation needed to reduce perinatal health inequities. Population health management programs should carefully consider algorithmic fairness in predictive models and associated resource allocation decisions.
期刊介绍:
Journal of Racial and Ethnic Health Disparities reports on the scholarly progress of work to understand, address, and ultimately eliminate health disparities based on race and ethnicity. Efforts to explore underlying causes of health disparities and to describe interventions that have been undertaken to address racial and ethnic health disparities are featured. Promising studies that are ongoing or studies that have longer term data are welcome, as are studies that serve as lessons for best practices in eliminating health disparities. Original research, systematic reviews, and commentaries presenting the state-of-the-art thinking on problems centered on health disparities will be considered for publication. We particularly encourage review articles that generate innovative and testable ideas, and constructive discussions and/or critiques of health disparities.Because the Journal of Racial and Ethnic Health Disparities receives a large number of submissions, about 30% of submissions to the Journal are sent out for full peer review.