Sanjiv Ranjan Das, Michele Donini, J. Gelman, Kevin Haas, Mila Hardt, Jared Katzman, K. Kenthapadi, Pedro Larroy, Pinar Yilmaz, Bilal Zafar
{"title":"金融中机器学习的公平性措施","authors":"Sanjiv Ranjan Das, Michele Donini, J. Gelman, Kevin Haas, Mila Hardt, Jared Katzman, K. Kenthapadi, Pedro Larroy, Pinar Yilmaz, Bilal Zafar","doi":"10.3905/jfds.2021.1.075","DOIUrl":null,"url":null,"abstract":"The authors present a machine learning pipeline for fairness-aware machine learning (FAML) in finance that encompasses metrics for fairness (and accuracy). Whereas accuracy metrics are well understood and the principal ones are used frequently, there is no consensus as to which of several available measures for fairness should be used in a generic manner in the financial services industry. The authors explore these measures and discuss which ones to focus on at various stages in the ML pipeline, pre-training and post-training, and they examine simple bias mitigation approaches. Using a standard dataset, they show that the sequencing in their FAML pipeline offers a cogent approach to arriving at a fair and accurate ML model. The authors discuss the intersection of bias metrics with legal considerations in the United States, and the entanglement of explainability and fairness is exemplified in the case study. They discuss possible approaches for training ML models while satisfying constraints imposed from various fairness metrics and the role of causality in assessing fairness. Key Findings ▪ Sources of bias are presented and a range of metrics is considered for machine learning applications in finance, both pre-training and post-training of models. ▪ A process of using the metrics to arrive at fair models is discussed. ▪ Various considerations for the choice of specific metrics are also analyzed.","PeriodicalId":199045,"journal":{"name":"The Journal of Financial Data Science","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"20","resultStr":"{\"title\":\"Fairness Measures for Machine Learning in Finance\",\"authors\":\"Sanjiv Ranjan Das, Michele Donini, J. Gelman, Kevin Haas, Mila Hardt, Jared Katzman, K. Kenthapadi, Pedro Larroy, Pinar Yilmaz, Bilal Zafar\",\"doi\":\"10.3905/jfds.2021.1.075\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The authors present a machine learning pipeline for fairness-aware machine learning (FAML) in finance that encompasses metrics for fairness (and accuracy). Whereas accuracy metrics are well understood and the principal ones are used frequently, there is no consensus as to which of several available measures for fairness should be used in a generic manner in the financial services industry. The authors explore these measures and discuss which ones to focus on at various stages in the ML pipeline, pre-training and post-training, and they examine simple bias mitigation approaches. Using a standard dataset, they show that the sequencing in their FAML pipeline offers a cogent approach to arriving at a fair and accurate ML model. The authors discuss the intersection of bias metrics with legal considerations in the United States, and the entanglement of explainability and fairness is exemplified in the case study. They discuss possible approaches for training ML models while satisfying constraints imposed from various fairness metrics and the role of causality in assessing fairness. Key Findings ▪ Sources of bias are presented and a range of metrics is considered for machine learning applications in finance, both pre-training and post-training of models. ▪ A process of using the metrics to arrive at fair models is discussed. ▪ Various considerations for the choice of specific metrics are also analyzed.\",\"PeriodicalId\":199045,\"journal\":{\"name\":\"The Journal of Financial Data Science\",\"volume\":\"37 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-09-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"20\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Journal of Financial Data Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3905/jfds.2021.1.075\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Journal of Financial Data Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3905/jfds.2021.1.075","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The authors present a machine learning pipeline for fairness-aware machine learning (FAML) in finance that encompasses metrics for fairness (and accuracy). Whereas accuracy metrics are well understood and the principal ones are used frequently, there is no consensus as to which of several available measures for fairness should be used in a generic manner in the financial services industry. The authors explore these measures and discuss which ones to focus on at various stages in the ML pipeline, pre-training and post-training, and they examine simple bias mitigation approaches. Using a standard dataset, they show that the sequencing in their FAML pipeline offers a cogent approach to arriving at a fair and accurate ML model. The authors discuss the intersection of bias metrics with legal considerations in the United States, and the entanglement of explainability and fairness is exemplified in the case study. They discuss possible approaches for training ML models while satisfying constraints imposed from various fairness metrics and the role of causality in assessing fairness. Key Findings ▪ Sources of bias are presented and a range of metrics is considered for machine learning applications in finance, both pre-training and post-training of models. ▪ A process of using the metrics to arrive at fair models is discussed. ▪ Various considerations for the choice of specific metrics are also analyzed.