Lucas Monteiro Paes, Ananda Theertha Suresh, Alex Beutel, Flavio P. Calmon, Ahmad Beirami
{"title":"通过条件风险价值测试进行多组公平性评估","authors":"Lucas Monteiro Paes, Ananda Theertha Suresh, Alex Beutel, Flavio P. Calmon, Ahmad Beirami","doi":"arxiv-2312.03867","DOIUrl":null,"url":null,"abstract":"Machine learning (ML) models used in prediction and classification tasks may\ndisplay performance disparities across population groups determined by\nsensitive attributes (e.g., race, sex, age). We consider the problem of\nevaluating the performance of a fixed ML model across population groups defined\nby multiple sensitive attributes (e.g., race and sex and age). Here, the sample\ncomplexity for estimating the worst-case performance gap across groups (e.g.,\nthe largest difference in error rates) increases exponentially with the number\nof group-denoting sensitive attributes. To address this issue, we propose an\napproach to test for performance disparities based on Conditional Value-at-Risk\n(CVaR). By allowing a small probabilistic slack on the groups over which a\nmodel has approximately equal performance, we show that the sample complexity\nrequired for discovering performance violations is reduced exponentially to be\nat most upper bounded by the square root of the number of groups. As a\nbyproduct of our analysis, when the groups are weighted by a specific prior\ndistribution, we show that R\\'enyi entropy of order $2/3$ of the prior\ndistribution captures the sample complexity of the proposed CVaR test\nalgorithm. Finally, we also show that there exists a non-i.i.d. data collection\nstrategy that results in a sample complexity independent of the number of\ngroups.","PeriodicalId":501433,"journal":{"name":"arXiv - CS - Information Theory","volume":"103 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-Group Fairness Evaluation via Conditional Value-at-Risk Testing\",\"authors\":\"Lucas Monteiro Paes, Ananda Theertha Suresh, Alex Beutel, Flavio P. Calmon, Ahmad Beirami\",\"doi\":\"arxiv-2312.03867\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Machine learning (ML) models used in prediction and classification tasks may\\ndisplay performance disparities across population groups determined by\\nsensitive attributes (e.g., race, sex, age). We consider the problem of\\nevaluating the performance of a fixed ML model across population groups defined\\nby multiple sensitive attributes (e.g., race and sex and age). Here, the sample\\ncomplexity for estimating the worst-case performance gap across groups (e.g.,\\nthe largest difference in error rates) increases exponentially with the number\\nof group-denoting sensitive attributes. To address this issue, we propose an\\napproach to test for performance disparities based on Conditional Value-at-Risk\\n(CVaR). By allowing a small probabilistic slack on the groups over which a\\nmodel has approximately equal performance, we show that the sample complexity\\nrequired for discovering performance violations is reduced exponentially to be\\nat most upper bounded by the square root of the number of groups. As a\\nbyproduct of our analysis, when the groups are weighted by a specific prior\\ndistribution, we show that R\\\\'enyi entropy of order $2/3$ of the prior\\ndistribution captures the sample complexity of the proposed CVaR test\\nalgorithm. Finally, we also show that there exists a non-i.i.d. data collection\\nstrategy that results in a sample complexity independent of the number of\\ngroups.\",\"PeriodicalId\":501433,\"journal\":{\"name\":\"arXiv - CS - Information Theory\",\"volume\":\"103 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-12-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Theory\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2312.03867\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2312.03867","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-Group Fairness Evaluation via Conditional Value-at-Risk Testing
Machine learning (ML) models used in prediction and classification tasks may
display performance disparities across population groups determined by
sensitive attributes (e.g., race, sex, age). We consider the problem of
evaluating the performance of a fixed ML model across population groups defined
by multiple sensitive attributes (e.g., race and sex and age). Here, the sample
complexity for estimating the worst-case performance gap across groups (e.g.,
the largest difference in error rates) increases exponentially with the number
of group-denoting sensitive attributes. To address this issue, we propose an
approach to test for performance disparities based on Conditional Value-at-Risk
(CVaR). By allowing a small probabilistic slack on the groups over which a
model has approximately equal performance, we show that the sample complexity
required for discovering performance violations is reduced exponentially to be
at most upper bounded by the square root of the number of groups. As a
byproduct of our analysis, when the groups are weighted by a specific prior
distribution, we show that R\'enyi entropy of order $2/3$ of the prior
distribution captures the sample complexity of the proposed CVaR test
algorithm. Finally, we also show that there exists a non-i.i.d. data collection
strategy that results in a sample complexity independent of the number of
groups.