首页 > 最新文献

Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society最新文献

英文 中文
Persistent Anti-Muslim Bias in Large Language Models 大型语言模型中持续的反穆斯林偏见
Pub Date : 2021-01-14 DOI: 10.1145/3461702.3462624
Abubakar Abid, Maheen Farooqi, James Y. Zou
It has been observed that large-scale language models capture undesirable societal biases, e.g. relating to race and gender; yet religious bias has been relatively unexplored. We demonstrate that GPT-3, a state-of-the-art contextual language model, captures persistent Muslim-violence bias. We probe GPT-3 in various ways, including prompt completion, analogical reasoning, and story generation, to understand this anti-Muslim bias, demonstrating that it appears consistently and creatively in different uses of the model and that it is severe even compared to biases about other religious groups. For instance, Muslim is analogized to terrorist in 23% of test cases, while Jewish is mapped to its most common stereotype, money, in 5% of test cases. We quantify the positive distraction needed to overcome this bias with adversarial text prompts, and find that use of the most positive 6 adjectives reduces violent completions for Muslims from 66% to 20%, but which is still higher than for other religious groups.
据观察,大规模的语言模型捕获了不受欢迎的社会偏见,例如与种族和性别有关的偏见;然而,宗教偏见相对来说还没有被研究过。我们证明了GPT-3,一个最先进的语境语言模型,捕捉到持续的穆斯林暴力偏见。我们以各种方式探究GPT-3,包括提示完成、类比推理和故事生成,以理解这种反穆斯林偏见,证明它在模型的不同使用中始终如一地、创造性地出现,甚至与对其他宗教团体的偏见相比,它也是严重的。例如,在23%的测试案例中,穆斯林被类比为恐怖分子,而在5%的测试案例中,犹太人被类比为最常见的刻板印象——金钱。我们量化了对抗文本提示克服这种偏见所需的积极分心,发现使用最积极的6个形容词将穆斯林的暴力完成率从66%降低到20%,但仍高于其他宗教群体。
{"title":"Persistent Anti-Muslim Bias in Large Language Models","authors":"Abubakar Abid, Maheen Farooqi, James Y. Zou","doi":"10.1145/3461702.3462624","DOIUrl":"https://doi.org/10.1145/3461702.3462624","url":null,"abstract":"It has been observed that large-scale language models capture undesirable societal biases, e.g. relating to race and gender; yet religious bias has been relatively unexplored. We demonstrate that GPT-3, a state-of-the-art contextual language model, captures persistent Muslim-violence bias. We probe GPT-3 in various ways, including prompt completion, analogical reasoning, and story generation, to understand this anti-Muslim bias, demonstrating that it appears consistently and creatively in different uses of the model and that it is severe even compared to biases about other religious groups. For instance, Muslim is analogized to terrorist in 23% of test cases, while Jewish is mapped to its most common stereotype, money, in 5% of test cases. We quantify the positive distraction needed to overcome this bias with adversarial text prompts, and find that use of the most positive 6 adjectives reduces violent completions for Muslims from 66% to 20%, but which is still higher than for other religious groups.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130306437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 252
Explainable AI and Adoption of Financial Algorithmic Advisors: An Experimental Study 可解释的人工智能和金融算法顾问的采用:一项实验研究
Pub Date : 2021-01-05 DOI: 10.1145/3461702.3462565
D. David, Yehezkel S. Resheff, Talia Tron
We study whether receiving advice from either a human or algorithmic advisor, accompanied by five types of Local and Global explanation labelings, has an effect on the readiness to adopt, willingness to pay, and trust in a financial AI consultant. We compare the differences over time and in various key situations using a unique experimental framework where participants play a web-based game with real monetary consequences. We observed that accuracy-based explanations of the model in initial phases leads to higher adoption rates. When the performance of the model is immaculate, there is less importance associated with the kind of explanation for adoption. Using more elaborate feature-based or accuracy-based explanations helps substantially in reducing the adoption drop upon model failure. Furthermore, using an autopilot increases adoption significantly. Participants assigned to the AI-labeled advice with explanations were willing to pay more for the advice than the AI-labeled advice with "No-explanation" alternative. These results add to the literature on the importance of XAI for algorithmic adoption and trust.
我们研究了接受人类或算法顾问的建议,以及五种类型的本地和全球解释标签,是否会对金融人工智能顾问的接受意愿、支付意愿和信任产生影响。我们使用一个独特的实验框架来比较不同时间和不同关键情况下的差异,在这个实验框架中,参与者玩一款带有真实金钱后果的网络游戏。我们观察到,在初始阶段基于准确性的模型解释导致更高的采用率。当模型的性能完美无缺时,对采用的解释就不那么重要了。使用更详细的基于特征或基于准确性的解释,有助于大大减少模型失败时的采用率下降。此外,使用自动驾驶仪可以显著提高采用率。被分配到人工智能标记的有解释的建议的参与者愿意为这些建议支付更多的钱,而不是人工智能标记的有“不解释”的建议。这些结果增加了关于XAI对算法采用和信任的重要性的文献。
{"title":"Explainable AI and Adoption of Financial Algorithmic Advisors: An Experimental Study","authors":"D. David, Yehezkel S. Resheff, Talia Tron","doi":"10.1145/3461702.3462565","DOIUrl":"https://doi.org/10.1145/3461702.3462565","url":null,"abstract":"We study whether receiving advice from either a human or algorithmic advisor, accompanied by five types of Local and Global explanation labelings, has an effect on the readiness to adopt, willingness to pay, and trust in a financial AI consultant. We compare the differences over time and in various key situations using a unique experimental framework where participants play a web-based game with real monetary consequences. We observed that accuracy-based explanations of the model in initial phases leads to higher adoption rates. When the performance of the model is immaculate, there is less importance associated with the kind of explanation for adoption. Using more elaborate feature-based or accuracy-based explanations helps substantially in reducing the adoption drop upon model failure. Furthermore, using an autopilot increases adoption significantly. Participants assigned to the AI-labeled advice with explanations were willing to pay more for the advice than the AI-labeled advice with \"No-explanation\" alternative. These results add to the literature on the importance of XAI for algorithmic adoption and trust.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125529014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
GAEA: Graph Augmentation for Equitable Access via Reinforcement Learning GAEA:通过强化学习实现公平访问的图增强
Pub Date : 2020-12-07 DOI: 10.1145/3461702.3462615
Govardana Sachithanandam Ramachandran, Ivan Brugere, L. Varshney, Caiming Xiong
Disparate access to resources by different subpopulations is a prevalent issue in societal and sociotechnical networks. For example, urban infrastructure networks may enable certain racial groups to more easily access resources such as high-quality schools, grocery stores, and polling places. Similarly, social networks within universities and organizations may enable certain groups to more easily access people with valuable information or influence. Here we introduce a new class of problems, Graph Augmentation for Equitable Access (GAEA), to enhance equity in networked systems by editing graph edges under budget constraints. We prove such problems are NP-hard, and cannot be approximated within a factor of (1-1/3e). We develop a principled, sample- and time- efficient Markov Reward Process (MRP)-based mechanism design framework for GAEA. Our algorithm outperforms baselines on a diverse set of synthetic graphs. We further demonstrate the method on real-world networks, by merging public census, school, and transportation datasets for the city of Chicago and applying our algorithm to find human-interpretable edits to the bus network that enhance equitable access to high-quality schools across racial groups. Further experiments on Facebook networks of universities yield sets of new social connections that would increase equitable access to certain attributed nodes across gender groups.
不同亚群对资源的不同获取是社会和社会技术网络中普遍存在的问题。例如,城市基础设施网络可以使某些种族群体更容易获得诸如高质量学校、杂货店和投票站等资源。同样,大学和组织内部的社会网络可能使某些团体更容易接触到有价值信息或有影响力的人。在这里,我们引入了一类新的问题,即公平访问的图增强(GAEA),通过在预算约束下编辑图边来增强网络系统的公平性。我们证明了这些问题是np困难的,并且不能在因子(1-1/3e)内近似。我们为GAEA开发了一个原则性、样本和时间效率高的基于马尔可夫奖励过程(MRP)的机制设计框架。我们的算法在一组不同的合成图上优于基线。通过合并芝加哥市的公共人口普查、学校和交通数据集,我们进一步在现实世界的网络上演示了该方法,并应用我们的算法找到了人类可解释的公交网络编辑,从而提高了跨种族群体获得高质量学校的公平机会。在Facebook大学网络上进行的进一步实验产生了一系列新的社会联系,这些联系将增加跨性别群体对某些属性节点的公平访问。
{"title":"GAEA: Graph Augmentation for Equitable Access via Reinforcement Learning","authors":"Govardana Sachithanandam Ramachandran, Ivan Brugere, L. Varshney, Caiming Xiong","doi":"10.1145/3461702.3462615","DOIUrl":"https://doi.org/10.1145/3461702.3462615","url":null,"abstract":"Disparate access to resources by different subpopulations is a prevalent issue in societal and sociotechnical networks. For example, urban infrastructure networks may enable certain racial groups to more easily access resources such as high-quality schools, grocery stores, and polling places. Similarly, social networks within universities and organizations may enable certain groups to more easily access people with valuable information or influence. Here we introduce a new class of problems, Graph Augmentation for Equitable Access (GAEA), to enhance equity in networked systems by editing graph edges under budget constraints. We prove such problems are NP-hard, and cannot be approximated within a factor of (1-1/3e). We develop a principled, sample- and time- efficient Markov Reward Process (MRP)-based mechanism design framework for GAEA. Our algorithm outperforms baselines on a diverse set of synthetic graphs. We further demonstrate the method on real-world networks, by merging public census, school, and transportation datasets for the city of Chicago and applying our algorithm to find human-interpretable edits to the bus network that enhance equitable access to high-quality schools across racial groups. Further experiments on Facebook networks of universities yield sets of new social connections that would increase equitable access to certain attributed nodes across gender groups.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132160052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
FairOD: Fairness-aware Outlier Detection 公平感知异常值检测
Pub Date : 2020-12-05 DOI: 10.1145/3461702.3462517
Shubhranshu Shekhar, Neil Shah, L. Akoglu
Fairness and Outlier Detection (OD) are closely related, as it is exactly the goal of OD to spot rare, minority samples in a given population. However, when being a minority (as defined by protected variables, such as race/ethnicity/sex/age) does not reflect positive-class membership (such as criminal/fraud), OD produces unjust outcomes. Surprisingly, fairness-aware OD has been almost untouched in prior work, as fair machine learning literature mainly focuses on supervised settings. Our work aims to bridge this gap. Specifically, we develop desiderata capturing well-motivated fairness criteria for OD, and systematically formalize the fair OD problem. Further, guided by our desiderata, we propose FairOD, a fairness-aware outlier detector that has the following desirable properties: FairOD (1) exhibits treatment parity at test time, (2) aims to flag equal proportions of samples from all groups (i.e. obtain group fairness, via statistical parity), and (3) strives to flag truly high-risk samples within each group. Extensive experiments on a diverse set of synthetic and real world datasets show that FairOD produces outcomes that are fair with respect to protected variables, while performing comparable to (and in some cases, even better than) fairness-agnostic detectors in terms of detection performance.
公平和离群值检测(OD)密切相关,因为OD的目标就是在给定的群体中发现罕见的少数样本。然而,当作为少数群体(由受保护的变量定义,如种族/民族/性别/年龄)不能反映积极的阶级成员(如犯罪/欺诈)时,OD会产生不公正的结果。令人惊讶的是,公平感知OD在之前的工作中几乎没有受到影响,因为公平机器学习文献主要关注监督设置。我们的工作旨在弥合这一差距。具体而言,我们开发了捕获动机良好的OD公平标准的期望,并系统地形式化了公平OD问题。此外,在我们期望的指导下,我们提出了FairOD,这是一个具有公平性意识的异常值检测器,具有以下理想特性:FairOD(1)在测试时显示处理平价,(2)旨在标记所有组中相同比例的样本(即通过统计平价获得组公平性),以及(3)努力标记每个组中真正的高风险样本。在不同的合成数据集和真实世界数据集上进行的大量实验表明,FairOD产生的结果对于受保护的变量是公平的,同时在检测性能方面与不公平的检测器相当(在某些情况下甚至更好)。
{"title":"FairOD: Fairness-aware Outlier Detection","authors":"Shubhranshu Shekhar, Neil Shah, L. Akoglu","doi":"10.1145/3461702.3462517","DOIUrl":"https://doi.org/10.1145/3461702.3462517","url":null,"abstract":"Fairness and Outlier Detection (OD) are closely related, as it is exactly the goal of OD to spot rare, minority samples in a given population. However, when being a minority (as defined by protected variables, such as race/ethnicity/sex/age) does not reflect positive-class membership (such as criminal/fraud), OD produces unjust outcomes. Surprisingly, fairness-aware OD has been almost untouched in prior work, as fair machine learning literature mainly focuses on supervised settings. Our work aims to bridge this gap. Specifically, we develop desiderata capturing well-motivated fairness criteria for OD, and systematically formalize the fair OD problem. Further, guided by our desiderata, we propose FairOD, a fairness-aware outlier detector that has the following desirable properties: FairOD (1) exhibits treatment parity at test time, (2) aims to flag equal proportions of samples from all groups (i.e. obtain group fairness, via statistical parity), and (3) strives to flag truly high-risk samples within each group. Extensive experiments on a diverse set of synthetic and real world datasets show that FairOD produces outcomes that are fair with respect to protected variables, while performing comparable to (and in some cases, even better than) fairness-agnostic detectors in terms of detection performance.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115275752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Does Fair Ranking Improve Minority Outcomes? Understanding the Interplay of Human and Algorithmic Biases in Online Hiring 公平排名能提高少数族裔的成绩吗?理解在线招聘中人类和算法偏见的相互作用
Pub Date : 2020-12-01 DOI: 10.1145/3461702.3462602
Tom Sühr, Sophie Hilgard, Himabindu Lakkaraju
Ranking algorithms are being widely employed in various online hiring platforms including LinkedIn, TaskRabbit, and Fiverr. Prior research has demonstrated that ranking algorithms employed by these platforms are prone to a variety of undesirable biases, leading to the proposal of fair ranking algorithms (e.g., Det-Greedy) which increase exposure of underrepresented candidates. However, there is little to no work that explores whether fair ranking algorithms actually improve real world outcomes (e.g., hiring decisions) for underrepresented groups. Furthermore, there is no clear understanding as to how other factors (e.g., job context, inherent biases of the employers) may impact the efficacy of fair ranking in practice. In this work, we analyze various sources of gender biases in online hiring platforms, including the job context and inherent biases of employers and establish how these factors interact with ranking algorithms to affect hiring decisions. To the best of our knowledge, this work makes the first attempt at studying the interplay between the aforementioned factors in the context of online hiring. We carry out a large-scale user study simulating online hiring scenarios with data from TaskRabbit, a popular online freelancing site. Our results demonstrate that while fair ranking algorithms generally improve the selection rates of underrepresented minorities, their effectiveness relies heavily on the job contexts and candidate profiles.
排名算法被广泛应用于各种在线招聘平台,包括LinkedIn、TaskRabbit和Fiverr。先前的研究表明,这些平台使用的排名算法容易产生各种不良偏见,导致提出公平排名算法(例如,Det-Greedy),这增加了代表性不足的候选人的曝光率。然而,对于代表性不足的群体来说,公平排名算法是否真的能改善现实世界的结果(例如招聘决策),几乎没有研究。此外,对于其他因素(如工作环境、雇主的固有偏见)在实践中如何影响公平排名的效果,目前还没有明确的认识。在这项工作中,我们分析了在线招聘平台中性别偏见的各种来源,包括工作环境和雇主的固有偏见,并建立了这些因素如何与排名算法相互作用以影响招聘决策。据我们所知,这项工作首次尝试研究在线招聘背景下上述因素之间的相互作用。我们进行了一项大规模的用户研究,利用TaskRabbit(一个流行的在线自由职业网站)的数据模拟在线招聘场景。我们的研究结果表明,虽然公平排名算法通常会提高代表性不足的少数族裔的选选率,但其有效性在很大程度上取决于工作背景和候选人简介。
{"title":"Does Fair Ranking Improve Minority Outcomes? Understanding the Interplay of Human and Algorithmic Biases in Online Hiring","authors":"Tom Sühr, Sophie Hilgard, Himabindu Lakkaraju","doi":"10.1145/3461702.3462602","DOIUrl":"https://doi.org/10.1145/3461702.3462602","url":null,"abstract":"Ranking algorithms are being widely employed in various online hiring platforms including LinkedIn, TaskRabbit, and Fiverr. Prior research has demonstrated that ranking algorithms employed by these platforms are prone to a variety of undesirable biases, leading to the proposal of fair ranking algorithms (e.g., Det-Greedy) which increase exposure of underrepresented candidates. However, there is little to no work that explores whether fair ranking algorithms actually improve real world outcomes (e.g., hiring decisions) for underrepresented groups. Furthermore, there is no clear understanding as to how other factors (e.g., job context, inherent biases of the employers) may impact the efficacy of fair ranking in practice. In this work, we analyze various sources of gender biases in online hiring platforms, including the job context and inherent biases of employers and establish how these factors interact with ranking algorithms to affect hiring decisions. To the best of our knowledge, this work makes the first attempt at studying the interplay between the aforementioned factors in the context of online hiring. We carry out a large-scale user study simulating online hiring scenarios with data from TaskRabbit, a popular online freelancing site. Our results demonstrate that while fair ranking algorithms generally improve the selection rates of underrepresented minorities, their effectiveness relies heavily on the job contexts and candidate profiles.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133320480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Uncertainty as a Form of Transparency: Measuring, Communicating, and Using Uncertainty 作为透明形式的不确定性:测量、沟通和使用不确定性
Pub Date : 2020-11-15 DOI: 10.1145/3461702.3462571
Umang Bhatt, Yunfeng Zhang, Javier Antorán, Q. Liao, P. Sattigeri, Riccardo Fogliato, Gabrielle Gauthier Melançon, R. Krishnan, Jason Stanley, Omesh Tickoo, L. Nachman, R. Chunara, Adrian Weller, Alice Xiang
Algorithmic transparency entails exposing system properties to various stakeholders for purposes that include understanding, improving, and contesting predictions. Until now, most research into algorithmic transparency has predominantly focused on explainability. Explainability attempts to provide reasons for a machine learning model's behavior to stakeholders. However, understanding a model's specific behavior alone might not be enough for stakeholders to gauge whether the model is wrong or lacks sufficient knowledge to solve the task at hand. In this paper, we argue for considering a complementary form of transparency by estimating and communicating the uncertainty associated with model predictions. First, we discuss methods for assessing uncertainty. Then, we characterize how uncertainty can be used to mitigate model unfairness, augment decision-making, and build trustworthy systems. Finally, we outline methods for displaying uncertainty to stakeholders and recommend how to collect information required for incorporating uncertainty into existing ML pipelines. This work constitutes an interdisciplinary review drawn from literature spanning machine learning, visualization/HCI, design, decision-making, and fairness. We aim to encourage researchers and practitioners to measure, communicate, and use uncertainty as a form of transparency.
算法透明性需要将系统属性暴露给各种涉众,以达到理解、改进和争论预测的目的。到目前为止,大多数关于算法透明性的研究主要集中在可解释性上。可解释性试图为机器学习模型对利益相关者的行为提供原因。然而,仅仅理解模型的特定行为可能不足以让涉众判断模型是否错误或缺乏足够的知识来解决手头的任务。在本文中,我们主张通过估计和传达与模型预测相关的不确定性来考虑透明度的补充形式。首先,我们讨论了评估不确定性的方法。然后,我们描述了如何使用不确定性来减轻模型不公平,增强决策,并建立可信赖的系统。最后,我们概述了向利益相关者显示不确定性的方法,并建议如何收集将不确定性纳入现有ML管道所需的信息。这项工作是一项跨学科的综述,从机器学习、可视化/人机交互、设计、决策和公平等方面的文献中提取。我们的目标是鼓励研究人员和从业人员测量、交流和使用不确定性作为一种透明度形式。
{"title":"Uncertainty as a Form of Transparency: Measuring, Communicating, and Using Uncertainty","authors":"Umang Bhatt, Yunfeng Zhang, Javier Antorán, Q. Liao, P. Sattigeri, Riccardo Fogliato, Gabrielle Gauthier Melançon, R. Krishnan, Jason Stanley, Omesh Tickoo, L. Nachman, R. Chunara, Adrian Weller, Alice Xiang","doi":"10.1145/3461702.3462571","DOIUrl":"https://doi.org/10.1145/3461702.3462571","url":null,"abstract":"Algorithmic transparency entails exposing system properties to various stakeholders for purposes that include understanding, improving, and contesting predictions. Until now, most research into algorithmic transparency has predominantly focused on explainability. Explainability attempts to provide reasons for a machine learning model's behavior to stakeholders. However, understanding a model's specific behavior alone might not be enough for stakeholders to gauge whether the model is wrong or lacks sufficient knowledge to solve the task at hand. In this paper, we argue for considering a complementary form of transparency by estimating and communicating the uncertainty associated with model predictions. First, we discuss methods for assessing uncertainty. Then, we characterize how uncertainty can be used to mitigate model unfairness, augment decision-making, and build trustworthy systems. Finally, we outline methods for displaying uncertainty to stakeholders and recommend how to collect information required for incorporating uncertainty into existing ML pipelines. This work constitutes an interdisciplinary review drawn from literature spanning machine learning, visualization/HCI, design, decision-making, and fairness. We aim to encourage researchers and practitioners to measure, communicate, and use uncertainty as a form of transparency.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126699620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 151
Towards Unifying Feature Attribution and Counterfactual Explanations: Different Means to the Same End 走向特征归因与反事实解释的统一:不同手段达到同一目的
Pub Date : 2020-11-10 DOI: 10.1145/3461702.3462597
R. Mothilal, Divyat Mahajan, Chenhao Tan, Amit Sharma
Feature attributions and counterfactual explanations are popular approaches to explain a ML model. The former assigns an importance score to each input feature, while the latter provides input examples with minimal changes to alter the model's predictions. To unify these approaches, we provide an interpretation based on the actual causality framework and present two key results in terms of their use. First, we present a method to generate feature attribution explanations from a set of counterfactual examples. These feature attributions convey how important a feature is to changing the classification outcome of a model, especially on whether a subset of features is necessary and/or sufficient for that change, which attribution-based methods are unable to provide. Second, we show how counterfactual examples can be used to evaluate the goodness of an attribution-based explanation in terms of its necessity and sufficiency. As a result, we highlight the complimentary of these two approaches. Our evaluation on three benchmark datasets --- Adult-Income, LendingClub, and German-Credit --- confirms the complimentary. Feature attribution methods like LIME and SHAP and counterfactual explanation methods like Wachter et al. and DiCE often do not agree on feature importance rankings. In addition, by restricting the features that can be modified for generating counterfactual examples, we find that the top-k features from LIME or SHAP are often neither necessary nor sufficient explanations of a model's prediction. Finally, we present a case study of different explanation methods on a real-world hospital triage problem.
特征归因和反事实解释是解释ML模型的常用方法。前者为每个输入特征分配一个重要分数,而后者提供最小变化的输入示例来改变模型的预测。为了统一这些方法,我们提供了一个基于实际因果关系框架的解释,并就其使用提出了两个关键结果。首先,我们提出了一种从一组反事实示例中生成特征归因解释的方法。这些特征属性传达了特征对于改变模型的分类结果有多重要,特别是关于特征子集对于该变化是否必要和/或充分,这是基于属性的方法无法提供的。其次,我们展示了如何使用反事实例子来评估基于归因的解释的必要性和充分性。因此,我们强调这两种方法的互补性。我们对三个基准数据集(Adult-Income, LendingClub和German-Credit)的评估证实了这一点。特征归因方法(如LIME和SHAP)和反事实解释方法(如Wachter et al.和DiCE)在特征重要性排名上往往不一致。此外,通过限制可以修改以生成反事实示例的特征,我们发现来自LIME或SHAP的top-k特征通常既不是模型预测的必要解释,也不是充分解释。最后,我们提出了一个案例研究不同的解释方法对现实世界的医院分诊问题。
{"title":"Towards Unifying Feature Attribution and Counterfactual Explanations: Different Means to the Same End","authors":"R. Mothilal, Divyat Mahajan, Chenhao Tan, Amit Sharma","doi":"10.1145/3461702.3462597","DOIUrl":"https://doi.org/10.1145/3461702.3462597","url":null,"abstract":"Feature attributions and counterfactual explanations are popular approaches to explain a ML model. The former assigns an importance score to each input feature, while the latter provides input examples with minimal changes to alter the model's predictions. To unify these approaches, we provide an interpretation based on the actual causality framework and present two key results in terms of their use. First, we present a method to generate feature attribution explanations from a set of counterfactual examples. These feature attributions convey how important a feature is to changing the classification outcome of a model, especially on whether a subset of features is necessary and/or sufficient for that change, which attribution-based methods are unable to provide. Second, we show how counterfactual examples can be used to evaluate the goodness of an attribution-based explanation in terms of its necessity and sufficiency. As a result, we highlight the complimentary of these two approaches. Our evaluation on three benchmark datasets --- Adult-Income, LendingClub, and German-Credit --- confirms the complimentary. Feature attribution methods like LIME and SHAP and counterfactual explanation methods like Wachter et al. and DiCE often do not agree on feature importance rankings. In addition, by restricting the features that can be modified for generating counterfactual examples, we find that the top-k features from LIME or SHAP are often neither necessary nor sufficient explanations of a model's prediction. Finally, we present a case study of different explanation methods on a real-world hospital triage problem.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116798208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
Fair Machine Learning Under Partial Compliance 部分遵从下的公平机器学习
Pub Date : 2020-11-07 DOI: 10.1145/3461702.3462521
Jessica Dai, S. Fazelpour, Zachary Chase Lipton
Typically, fair machine learning research focuses on a single decision maker and assumes that the underlying population is stationary. However, many of the critical domains motivating this work are characterized by competitive marketplaces with many decision makers. Realistically, we might expect only a subset of them to adopt any non-compulsory fairness-conscious policy, a situation that political philosophers call partial compliance. This possibility raises important questions: how does partial compliance and the consequent strategic behavior of decision subjects affect the allocation outcomes? If k% of employers were to voluntarily adopt a fairness-promoting intervention, should we expect k% progress (in aggregate) towards the benefits of universal adoption, or will the dynamics of partial compliance wash out the hoped-for benefits? How might adopting a global (versus local) perspective impact the conclusions of an auditor? In this paper, we propose a simple model of an employment market, leveraging simulation as a tool to explore the impact of both interaction effects and incentive effects on outcomes and auditing metrics. Our key findings are that at equilibrium: (1) partial compliance by k% of employers can result in far less than proportional (k%) progress towards the full compliance outcomes; (2) the gap is more severe when fair employers match global (vs local) statistics; (3) choices of local vs global statistics can paint dramatically different pictures of the performance vis-a-vis fairness desiderata of compliant versus non-compliant employers; (4) partial compliance based on local parity measures can induce extreme segregation. Finally, we discuss implications for auditors and insights concerning the design of regulatory frameworks.
通常,公平的机器学习研究只关注一个决策者,并假设潜在的人口是平稳的。然而,推动这项工作的许多关键领域的特点是具有许多决策者的竞争性市场。实际上,我们可能只期望他们中的一部分人采取任何非强制性的公平意识政策,这种情况被政治哲学家称为部分服从。这种可能性提出了重要的问题:决策主体的部分服从和随之而来的战略行为如何影响分配结果?如果k%的雇主自愿采取促进公平的干预措施,我们是否应该期待k%的进展(总的来说)朝着普遍采用的好处,或者部分遵守的动态会冲毁希望的好处?采用全局视角(相对于局部视角)会如何影响审核员的结论?在本文中,我们提出了一个简单的就业市场模型,利用模拟作为工具来探索互动效应和激励效应对结果和审计指标的影响。我们的主要发现是,在均衡状态下:(1)k%的雇主的部分合规可能导致远低于比例(k%)的进步,以实现完全合规的结果;(2)当公平雇主匹配全球(相对于本地)统计数据时,差距会更大;(3)本地和全球统计数据的选择可以描绘出合规雇主和不合规雇主的绩效与公平期望之间的巨大差异;(4)基于局部宇称测度的局部服从会导致极端偏析。最后,我们讨论了对审计师的影响以及对监管框架设计的见解。
{"title":"Fair Machine Learning Under Partial Compliance","authors":"Jessica Dai, S. Fazelpour, Zachary Chase Lipton","doi":"10.1145/3461702.3462521","DOIUrl":"https://doi.org/10.1145/3461702.3462521","url":null,"abstract":"Typically, fair machine learning research focuses on a single decision maker and assumes that the underlying population is stationary. However, many of the critical domains motivating this work are characterized by competitive marketplaces with many decision makers. Realistically, we might expect only a subset of them to adopt any non-compulsory fairness-conscious policy, a situation that political philosophers call partial compliance. This possibility raises important questions: how does partial compliance and the consequent strategic behavior of decision subjects affect the allocation outcomes? If k% of employers were to voluntarily adopt a fairness-promoting intervention, should we expect k% progress (in aggregate) towards the benefits of universal adoption, or will the dynamics of partial compliance wash out the hoped-for benefits? How might adopting a global (versus local) perspective impact the conclusions of an auditor? In this paper, we propose a simple model of an employment market, leveraging simulation as a tool to explore the impact of both interaction effects and incentive effects on outcomes and auditing metrics. Our key findings are that at equilibrium: (1) partial compliance by k% of employers can result in far less than proportional (k%) progress towards the full compliance outcomes; (2) the gap is more severe when fair employers match global (vs local) statistics; (3) choices of local vs global statistics can paint dramatically different pictures of the performance vis-a-vis fairness desiderata of compliant versus non-compliant employers; (4) partial compliance based on local parity measures can induce extreme segregation. Finally, we discuss implications for auditors and insights concerning the design of regulatory frameworks.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124633682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Causal Multi-level Fairness 因果多层次公平
Pub Date : 2020-10-14 DOI: 10.1145/3461702.3462587
V. Mhasawade, R. Chunara
Algorithmic systems are known to impact marginalized groups severely, and more so, if all sources of bias are not considered. While work in algorithmic fairness to-date has primarily focused on addressing discrimination due to individually linked attributes, social science research elucidates how some properties we link to individuals can be conceptualized as having causes at macro (e.g. structural) levels, and it may be important to be fair to attributes at multiple levels. For example, instead of simply considering race as a causal, protected attribute of an individual, the cause may be distilled as perceived racial discrimination an individual experiences, which in turn can be affected by neighborhood-level factors. This multi-level conceptualization is relevant to questions of fairness, as it may not only be important to take into account if the individual belonged to another demographic group, but also if the individual received advantaged treatment at the macro-level. In this paper, we formalize the problem of multi-level fairness using tools from causal inference in a manner that allows one to assess and account for effects of sensitive attributes at multiple levels. We show importance of the problem by illustrating residual unfairness if macro-level sensitive attributes are not accounted for, or included without accounting for their multi-level nature. Further, in the context of a real-world task of predicting income based on macro and individual-level attributes, we demonstrate an approach for mitigating unfairness, a result of multi-level sensitive attributes.
众所周知,算法系统会严重影响边缘群体,如果不考虑所有偏见来源,这种影响会更大。虽然到目前为止,算法公平的工作主要集中在解决由于个体相关属性造成的歧视,但社会科学研究阐明了我们与个体相关的一些属性如何被概念化为宏观(例如结构)层面的原因,并且在多个层面上对属性公平可能很重要。例如,与其简单地将种族视为个人的因果关系和受保护的属性,不如将原因提炼为个人经历的种族歧视,而种族歧视又可能受到社区因素的影响。这种多层次的概念化与公平问题有关,因为不仅要考虑到个人是否属于另一个人口群体,而且要考虑到个人是否在宏观层面上得到了有利待遇。在本文中,我们使用因果推理的工具形式化了多层次公平问题,这种方式允许人们评估和解释多层次敏感属性的影响。我们通过说明如果未考虑宏观级敏感属性或包含未考虑其多层次性质的剩余不公平性来显示问题的重要性。此外,在基于宏观和个人层面属性预测收入的现实世界任务的背景下,我们展示了一种减轻多层次敏感属性导致的不公平的方法。
{"title":"Causal Multi-level Fairness","authors":"V. Mhasawade, R. Chunara","doi":"10.1145/3461702.3462587","DOIUrl":"https://doi.org/10.1145/3461702.3462587","url":null,"abstract":"Algorithmic systems are known to impact marginalized groups severely, and more so, if all sources of bias are not considered. While work in algorithmic fairness to-date has primarily focused on addressing discrimination due to individually linked attributes, social science research elucidates how some properties we link to individuals can be conceptualized as having causes at macro (e.g. structural) levels, and it may be important to be fair to attributes at multiple levels. For example, instead of simply considering race as a causal, protected attribute of an individual, the cause may be distilled as perceived racial discrimination an individual experiences, which in turn can be affected by neighborhood-level factors. This multi-level conceptualization is relevant to questions of fairness, as it may not only be important to take into account if the individual belonged to another demographic group, but also if the individual received advantaged treatment at the macro-level. In this paper, we formalize the problem of multi-level fairness using tools from causal inference in a manner that allows one to assess and account for effects of sensitive attributes at multiple levels. We show importance of the problem by illustrating residual unfairness if macro-level sensitive attributes are not accounted for, or included without accounting for their multi-level nature. Further, in the context of a real-world task of predicting income based on macro and individual-level attributes, we demonstrate an approach for mitigating unfairness, a result of multi-level sensitive attributes.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134255873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
FaiR-N: Fair and Robust Neural Networks for Structured Data Fair - n:结构化数据的公平和鲁棒神经网络
Pub Date : 2020-10-13 DOI: 10.1145/3461702.3462559
Shubham Sharma, Alan H. Gee, D. Paydarfar, J. Ghosh
Fairness and robustness in machine learning are crucial when individuals are subject to automated decisions made by models in high-stake domains. To promote ethical artificial intelligence, fairness metrics that rely on comparing model error rates across subpopulations have been widely investigated for the detection and mitigation of bias. However, fairness measures that rely on comparing the ability to achieve recourse have been relatively unexplored. In this paper, we present a novel formulation for training neural networks that considers the distance of data observations to the decision boundary such that the new objective: (1) reduces the disparity in the average ability of recourse between individuals in each protected group, and (2) increases the average distance of data points to the boundary to promote adversarial robustness. We demonstrate that models trained with this new objective are more fair and adversarially robust neural networks, with similar accuracies, when compared to models without it. We also investigate a trade-off between the recourse-based fairness and robustness objectives. Moreover, we qualitatively motivate and empirically show that reducing recourse disparity across protected groups also improves fairness measures that rely on error rates. To the best of our knowledge, this is the first time that recourse disparity across groups are considered to train fairer neural networks.
当个人受制于高风险领域模型的自动决策时,机器学习中的公平性和鲁棒性至关重要。为了促进道德人工智能,人们已经广泛研究了依赖于比较不同亚群体的模型错误率的公平指标,以检测和减轻偏见。然而,相对而言,以比较获得追索权的能力为依据的公平措施尚未得到探索。在本文中,我们提出了一种新的训练神经网络的公式,该公式考虑了数据观测到决策边界的距离,从而实现了新的目标:(1)减少每个受保护群体中个体之间平均求助能力的差异;(2)增加数据点到边界的平均距离,以提高对抗鲁棒性。我们证明,与没有这个新目标的模型相比,用这个新目标训练的模型是更公平和对抗鲁棒的神经网络,具有相似的精度。我们还研究了基于资源的公平性和鲁棒性目标之间的权衡。此外,我们定性地激励和实证地表明,减少受保护群体之间的追索权差距也改善了依赖错误率的公平措施。据我们所知,这是第一次考虑群体之间的追索权差异来训练更公平的神经网络。
{"title":"FaiR-N: Fair and Robust Neural Networks for Structured Data","authors":"Shubham Sharma, Alan H. Gee, D. Paydarfar, J. Ghosh","doi":"10.1145/3461702.3462559","DOIUrl":"https://doi.org/10.1145/3461702.3462559","url":null,"abstract":"Fairness and robustness in machine learning are crucial when individuals are subject to automated decisions made by models in high-stake domains. To promote ethical artificial intelligence, fairness metrics that rely on comparing model error rates across subpopulations have been widely investigated for the detection and mitigation of bias. However, fairness measures that rely on comparing the ability to achieve recourse have been relatively unexplored. In this paper, we present a novel formulation for training neural networks that considers the distance of data observations to the decision boundary such that the new objective: (1) reduces the disparity in the average ability of recourse between individuals in each protected group, and (2) increases the average distance of data points to the boundary to promote adversarial robustness. We demonstrate that models trained with this new objective are more fair and adversarially robust neural networks, with similar accuracies, when compared to models without it. We also investigate a trade-off between the recourse-based fairness and robustness objectives. Moreover, we qualitatively motivate and empirically show that reducing recourse disparity across protected groups also improves fairness measures that rely on error rates. To the best of our knowledge, this is the first time that recourse disparity across groups are considered to train fairer neural networks.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128029325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1