{"title":"Mitigating Sex Bias in Audio Data-driven COPD and COVID-19 Breathing Pattern Detection Models","authors":"Rachel Pfeifer, Sudip Vhaduri, James Eric Dietz","doi":"arxiv-2409.10677","DOIUrl":null,"url":null,"abstract":"In the healthcare industry, researchers have been developing machine learning\nmodels to automate diagnosing patients with respiratory illnesses based on\ntheir breathing patterns. However, these models do not consider the demographic\nbiases, particularly sex bias, that often occur when models are trained with a\nskewed patient dataset. Hence, it is essential in such an important industry to\nreduce this bias so that models can make fair diagnoses. In this work, we\nexamine the bias in models used to detect breathing patterns of two major\nrespiratory diseases, i.e., chronic obstructive pulmonary disease (COPD) and\nCOVID-19. Using decision tree models trained with audio recordings of breathing\npatterns obtained from two open-source datasets consisting of 29 COPD and 680\nCOVID-19-positive patients, we analyze the effect of sex bias on the models.\nWith a threshold optimizer and two constraints (demographic parity and\nequalized odds) to mitigate the bias, we witness 81.43% (demographic parity\ndifference) and 71.81% (equalized odds difference) improvements. These findings\nare statistically significant.","PeriodicalId":501284,"journal":{"name":"arXiv - EE - Audio and Speech Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Audio and Speech Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10677","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In the healthcare industry, researchers have been developing machine learning
models to automate diagnosing patients with respiratory illnesses based on
their breathing patterns. However, these models do not consider the demographic
biases, particularly sex bias, that often occur when models are trained with a
skewed patient dataset. Hence, it is essential in such an important industry to
reduce this bias so that models can make fair diagnoses. In this work, we
examine the bias in models used to detect breathing patterns of two major
respiratory diseases, i.e., chronic obstructive pulmonary disease (COPD) and
COVID-19. Using decision tree models trained with audio recordings of breathing
patterns obtained from two open-source datasets consisting of 29 COPD and 680
COVID-19-positive patients, we analyze the effect of sex bias on the models.
With a threshold optimizer and two constraints (demographic parity and
equalized odds) to mitigate the bias, we witness 81.43% (demographic parity
difference) and 71.81% (equalized odds difference) improvements. These findings
are statistically significant.