Constantine Tarabanis MD , Evangelos Kalampokis PhD , Mahmoud Khalil MD , Carlos L. Alviar MD , Larry A. Chinitz MD, FHRS , Lior Jankelson MD, PhD
{"title":"心肌梗死后住院死亡率的可解释SHAP-XGBoost模型","authors":"Constantine Tarabanis MD , Evangelos Kalampokis PhD , Mahmoud Khalil MD , Carlos L. Alviar MD , Larry A. Chinitz MD, FHRS , Lior Jankelson MD, PhD","doi":"10.1016/j.cvdhj.2023.06.001","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><p>A lack of explainability in published machine learning (ML) models limits clinicians’ understanding of how predictions are made, in turn undermining uptake of the models into clinical practice.</p></div><div><h3>Objective</h3><p>The purpose of this study was to develop explainable ML models to predict in-hospital mortality in patients hospitalized for myocardial infarction (MI).</p></div><div><h3>Methods</h3><p>Adult patients hospitalized for an MI were identified in the National Inpatient Sample between January 1, 2012, and September 30, 2015. The resulting cohort comprised 457,096 patients described by 64 predictor variables relating to demographic/comorbidity characteristics and in-hospital complications. The gradient boosting algorithm eXtreme Gradient Boosting (XGBoost) was used to develop explainable models for in-hospital mortality prediction in the overall cohort and patient subgroups based on MI type and/or sex.</p></div><div><h3>Results</h3><p>The resulting models exhibited an area under the receiver operating characteristic curve (AUC) ranging from 0.876 to 0.942, specificity 82% to 87%, and sensitivity 75% to 87%. All models exhibited high negative predictive value ≥0.974. The SHapley Additive exPlanation (SHAP) framework was applied to explain the models. The top predictor variables of increasing and decreasing mortality were age and undergoing percutaneous coronary intervention, respectively. Other notable findings included a decreased mortality risk associated with certain patient subpopulations with hyperlipidemia and a comparatively greater risk of death among women below age 55 years.</p></div><div><h3>Conclusion</h3><p>The literature lacks explainable ML models predicting in-hospital mortality after an MI. In a national registry, explainable ML models performed best in ruling out in-hospital death post-MI, and their explanation illustrated their potential for guiding hypothesis generation and future study design.</p></div>","PeriodicalId":72527,"journal":{"name":"Cardiovascular digital health journal","volume":"4 4","pages":"Pages 126-132"},"PeriodicalIF":2.6000,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10435947/pdf/","citationCount":"0","resultStr":"{\"title\":\"Explainable SHAP-XGBoost models for in-hospital mortality after myocardial infarction\",\"authors\":\"Constantine Tarabanis MD , Evangelos Kalampokis PhD , Mahmoud Khalil MD , Carlos L. Alviar MD , Larry A. Chinitz MD, FHRS , Lior Jankelson MD, PhD\",\"doi\":\"10.1016/j.cvdhj.2023.06.001\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Background</h3><p>A lack of explainability in published machine learning (ML) models limits clinicians’ understanding of how predictions are made, in turn undermining uptake of the models into clinical practice.</p></div><div><h3>Objective</h3><p>The purpose of this study was to develop explainable ML models to predict in-hospital mortality in patients hospitalized for myocardial infarction (MI).</p></div><div><h3>Methods</h3><p>Adult patients hospitalized for an MI were identified in the National Inpatient Sample between January 1, 2012, and September 30, 2015. The resulting cohort comprised 457,096 patients described by 64 predictor variables relating to demographic/comorbidity characteristics and in-hospital complications. The gradient boosting algorithm eXtreme Gradient Boosting (XGBoost) was used to develop explainable models for in-hospital mortality prediction in the overall cohort and patient subgroups based on MI type and/or sex.</p></div><div><h3>Results</h3><p>The resulting models exhibited an area under the receiver operating characteristic curve (AUC) ranging from 0.876 to 0.942, specificity 82% to 87%, and sensitivity 75% to 87%. All models exhibited high negative predictive value ≥0.974. The SHapley Additive exPlanation (SHAP) framework was applied to explain the models. The top predictor variables of increasing and decreasing mortality were age and undergoing percutaneous coronary intervention, respectively. Other notable findings included a decreased mortality risk associated with certain patient subpopulations with hyperlipidemia and a comparatively greater risk of death among women below age 55 years.</p></div><div><h3>Conclusion</h3><p>The literature lacks explainable ML models predicting in-hospital mortality after an MI. In a national registry, explainable ML models performed best in ruling out in-hospital death post-MI, and their explanation illustrated their potential for guiding hypothesis generation and future study design.</p></div>\",\"PeriodicalId\":72527,\"journal\":{\"name\":\"Cardiovascular digital health journal\",\"volume\":\"4 4\",\"pages\":\"Pages 126-132\"},\"PeriodicalIF\":2.6000,\"publicationDate\":\"2023-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10435947/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Cardiovascular digital health journal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2666693623000361\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"CARDIAC & CARDIOVASCULAR SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cardiovascular digital health journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666693623000361","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"CARDIAC & CARDIOVASCULAR SYSTEMS","Score":null,"Total":0}
Explainable SHAP-XGBoost models for in-hospital mortality after myocardial infarction
Background
A lack of explainability in published machine learning (ML) models limits clinicians’ understanding of how predictions are made, in turn undermining uptake of the models into clinical practice.
Objective
The purpose of this study was to develop explainable ML models to predict in-hospital mortality in patients hospitalized for myocardial infarction (MI).
Methods
Adult patients hospitalized for an MI were identified in the National Inpatient Sample between January 1, 2012, and September 30, 2015. The resulting cohort comprised 457,096 patients described by 64 predictor variables relating to demographic/comorbidity characteristics and in-hospital complications. The gradient boosting algorithm eXtreme Gradient Boosting (XGBoost) was used to develop explainable models for in-hospital mortality prediction in the overall cohort and patient subgroups based on MI type and/or sex.
Results
The resulting models exhibited an area under the receiver operating characteristic curve (AUC) ranging from 0.876 to 0.942, specificity 82% to 87%, and sensitivity 75% to 87%. All models exhibited high negative predictive value ≥0.974. The SHapley Additive exPlanation (SHAP) framework was applied to explain the models. The top predictor variables of increasing and decreasing mortality were age and undergoing percutaneous coronary intervention, respectively. Other notable findings included a decreased mortality risk associated with certain patient subpopulations with hyperlipidemia and a comparatively greater risk of death among women below age 55 years.
Conclusion
The literature lacks explainable ML models predicting in-hospital mortality after an MI. In a national registry, explainable ML models performed best in ruling out in-hospital death post-MI, and their explanation illustrated their potential for guiding hypothesis generation and future study design.