{"title":"减轻音频数据驱动的慢性阻塞性肺病和 COVID-19 呼吸模式检测模型中的性别偏差","authors":"Rachel Pfeifer, Sudip Vhaduri, James Eric Dietz","doi":"arxiv-2409.10677","DOIUrl":null,"url":null,"abstract":"In the healthcare industry, researchers have been developing machine learning\nmodels to automate diagnosing patients with respiratory illnesses based on\ntheir breathing patterns. However, these models do not consider the demographic\nbiases, particularly sex bias, that often occur when models are trained with a\nskewed patient dataset. Hence, it is essential in such an important industry to\nreduce this bias so that models can make fair diagnoses. In this work, we\nexamine the bias in models used to detect breathing patterns of two major\nrespiratory diseases, i.e., chronic obstructive pulmonary disease (COPD) and\nCOVID-19. Using decision tree models trained with audio recordings of breathing\npatterns obtained from two open-source datasets consisting of 29 COPD and 680\nCOVID-19-positive patients, we analyze the effect of sex bias on the models.\nWith a threshold optimizer and two constraints (demographic parity and\nequalized odds) to mitigate the bias, we witness 81.43% (demographic parity\ndifference) and 71.81% (equalized odds difference) improvements. These findings\nare statistically significant.","PeriodicalId":501284,"journal":{"name":"arXiv - EE - Audio and Speech Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Mitigating Sex Bias in Audio Data-driven COPD and COVID-19 Breathing Pattern Detection Models\",\"authors\":\"Rachel Pfeifer, Sudip Vhaduri, James Eric Dietz\",\"doi\":\"arxiv-2409.10677\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the healthcare industry, researchers have been developing machine learning\\nmodels to automate diagnosing patients with respiratory illnesses based on\\ntheir breathing patterns. However, these models do not consider the demographic\\nbiases, particularly sex bias, that often occur when models are trained with a\\nskewed patient dataset. Hence, it is essential in such an important industry to\\nreduce this bias so that models can make fair diagnoses. In this work, we\\nexamine the bias in models used to detect breathing patterns of two major\\nrespiratory diseases, i.e., chronic obstructive pulmonary disease (COPD) and\\nCOVID-19. Using decision tree models trained with audio recordings of breathing\\npatterns obtained from two open-source datasets consisting of 29 COPD and 680\\nCOVID-19-positive patients, we analyze the effect of sex bias on the models.\\nWith a threshold optimizer and two constraints (demographic parity and\\nequalized odds) to mitigate the bias, we witness 81.43% (demographic parity\\ndifference) and 71.81% (equalized odds difference) improvements. These findings\\nare statistically significant.\",\"PeriodicalId\":501284,\"journal\":{\"name\":\"arXiv - EE - Audio and Speech Processing\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - EE - Audio and Speech Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10677\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Audio and Speech Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10677","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Mitigating Sex Bias in Audio Data-driven COPD and COVID-19 Breathing Pattern Detection Models
In the healthcare industry, researchers have been developing machine learning
models to automate diagnosing patients with respiratory illnesses based on
their breathing patterns. However, these models do not consider the demographic
biases, particularly sex bias, that often occur when models are trained with a
skewed patient dataset. Hence, it is essential in such an important industry to
reduce this bias so that models can make fair diagnoses. In this work, we
examine the bias in models used to detect breathing patterns of two major
respiratory diseases, i.e., chronic obstructive pulmonary disease (COPD) and
COVID-19. Using decision tree models trained with audio recordings of breathing
patterns obtained from two open-source datasets consisting of 29 COPD and 680
COVID-19-positive patients, we analyze the effect of sex bias on the models.
With a threshold optimizer and two constraints (demographic parity and
equalized odds) to mitigate the bias, we witness 81.43% (demographic parity
difference) and 71.81% (equalized odds difference) improvements. These findings
are statistically significant.