首页 > 最新文献

Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society最新文献

英文 中文
Understanding the Representation and Representativeness of Age in AI Data Sets 理解人工智能数据集中年龄的表征和代表性
Pub Date : 2021-03-10 DOI: 10.1145/3461702.3462590
J. Park, Michael S. Bernstein, Robin N. Brewer, Ece Kamar, M. Morris
A diverse representation of different demographic groups in AI training data sets is important in ensuring that the models will work for a large range of users. To this end, recent efforts in AI fairness and inclusion have advocated for creating AI data sets that are well-balanced across race, gender, socioeconomic status, and disability status. In this paper, we contribute to this line of work by focusing on the representation of age by asking whether older adults are represented proportionally to the population at large in AI data sets. We examine publicly-available information about 92 face data sets to understand how they codify age as a case study to investigate how the subjects' ages are recorded and whether older generations are represented. We find that older adults are very under-represented; five data sets in the study that explicitly documented the closed age intervals of their subjects included older adults (defined as older than 65 years), while only one included oldest-old adults (defined as older than 85 years). Additionally, we find that only 24 of the data sets include any age-related information in their documentation or metadata, and that there is no consistent method followed across these data sets to collect and record the subjects' ages. We recognize the unique difficulties in creating representative data sets in terms of age, but raise it as an important dimension that researchers and engineers interested in inclusive AI should consider.
在人工智能训练数据集中,不同人口群体的多样化表示对于确保模型适用于大范围的用户非常重要。为此,最近在人工智能公平性和包容性方面的努力提倡创建在种族、性别、社会经济地位和残疾状况之间保持良好平衡的人工智能数据集。在本文中,我们通过询问老年人在人工智能数据集中是否与总体人口成比例地代表年龄,从而关注年龄的表示,从而为这一工作做出贡献。我们研究了关于92个面部数据集的公开信息,以了解它们是如何将年龄编纂为案例研究,以调查受试者的年龄是如何记录的,以及是否代表了老一辈。我们发现老年人的比例非常低;研究中有五个数据集明确记录了受试者的封闭年龄间隔,其中包括老年人(定义为65岁以上),而只有一个数据集包括老年人(定义为85岁以上)。此外,我们发现只有24个数据集在其文档或元数据中包含任何与年龄相关的信息,并且这些数据集没有遵循一致的方法来收集和记录受试者的年龄。我们认识到在年龄方面创建代表性数据集的独特困难,但将其作为对包容性人工智能感兴趣的研究人员和工程师应该考虑的重要维度。
{"title":"Understanding the Representation and Representativeness of Age in AI Data Sets","authors":"J. Park, Michael S. Bernstein, Robin N. Brewer, Ece Kamar, M. Morris","doi":"10.1145/3461702.3462590","DOIUrl":"https://doi.org/10.1145/3461702.3462590","url":null,"abstract":"A diverse representation of different demographic groups in AI training data sets is important in ensuring that the models will work for a large range of users. To this end, recent efforts in AI fairness and inclusion have advocated for creating AI data sets that are well-balanced across race, gender, socioeconomic status, and disability status. In this paper, we contribute to this line of work by focusing on the representation of age by asking whether older adults are represented proportionally to the population at large in AI data sets. We examine publicly-available information about 92 face data sets to understand how they codify age as a case study to investigate how the subjects' ages are recorded and whether older generations are represented. We find that older adults are very under-represented; five data sets in the study that explicitly documented the closed age intervals of their subjects included older adults (defined as older than 65 years), while only one included oldest-old adults (defined as older than 85 years). Additionally, we find that only 24 of the data sets include any age-related information in their documentation or metadata, and that there is no consistent method followed across these data sets to collect and record the subjects' ages. We recognize the unique difficulties in creating representative data sets in terms of age, but raise it as an important dimension that researchers and engineers interested in inclusive AI should consider.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114142909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Designing Disaggregated Evaluations of AI Systems: Choices, Considerations, and Tradeoffs 设计人工智能系统的分类评估:选择、考虑和权衡
Pub Date : 2021-03-10 DOI: 10.1145/3461702.3462610
Solon Barocas, Anhong Guo, Ece Kamar, J. Krones, M. Morris, Jennifer Wortman Vaughan, Duncan Wadsworth, Hanna M. Wallach
Disaggregated evaluations of AI systems, in which system performance is assessed and reported separately for different groups of people, are conceptually simple. However, their design involves a variety of choices. Some of these choices influence the results that will be obtained, and thus the conclusions that can be drawn; others influence the impacts---both beneficial and harmful---that a disaggregated evaluation will have on people, including the people whose data is used to conduct the evaluation. We argue that a deeper understanding of these choices will enable researchers and practitioners to design careful and conclusive disaggregated evaluations. We also argue that better documentation of these choices, along with the underlying considerations and tradeoffs that have been made, will help others when interpreting an evaluation's results and conclusions.
人工智能系统的分类评估,即针对不同人群分别评估和报告系统性能,在概念上很简单。然而,它们的设计涉及多种选择。其中一些选择影响将获得的结果,从而影响可以得出的结论;其他因素会影响分类评估对人们(包括其数据被用于进行评估的人)产生的影响,既有有益的影响,也有有害的影响。我们认为,对这些选择的更深入的理解将使研究人员和从业者能够设计仔细和结论性的分类评估。我们还认为,更好地记录这些选择,以及所做的潜在考虑和权衡,将有助于其他人解释评估的结果和结论。
{"title":"Designing Disaggregated Evaluations of AI Systems: Choices, Considerations, and Tradeoffs","authors":"Solon Barocas, Anhong Guo, Ece Kamar, J. Krones, M. Morris, Jennifer Wortman Vaughan, Duncan Wadsworth, Hanna M. Wallach","doi":"10.1145/3461702.3462610","DOIUrl":"https://doi.org/10.1145/3461702.3462610","url":null,"abstract":"Disaggregated evaluations of AI systems, in which system performance is assessed and reported separately for different groups of people, are conceptually simple. However, their design involves a variety of choices. Some of these choices influence the results that will be obtained, and thus the conclusions that can be drawn; others influence the impacts---both beneficial and harmful---that a disaggregated evaluation will have on people, including the people whose data is used to conduct the evaluation. We argue that a deeper understanding of these choices will enable researchers and practitioners to design careful and conclusive disaggregated evaluations. We also argue that better documentation of these choices, along with the underlying considerations and tradeoffs that have been made, will help others when interpreting an evaluation's results and conclusions.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125046474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 53
Measuring Model Biases in the Absence of Ground Truth 在缺乏基础真理的情况下测量模型偏差
Pub Date : 2021-03-05 DOI: 10.1145/3461702.3462557
Osman Aka, Ken Burke, Alex Bauerle, Christina Greer, Margaret Mitchell
The measurement of bias in machine learning often focuses on model performance across identity subgroups (such as man and woman) with respect to groundtruth labels. However, these methods do not directly measure the associations that a model may have learned, for example between labels and identity subgroups. Further, measuring a model's bias requires a fully annotated evaluation dataset which may not be easily available in practice. We present an elegant mathematical solution that tackles both issues simultaneously, using image classification as a working example. By treating a classification model's predictions for a given image as a set of labels analogous to a "bag of words", we rank the biases that a model has learned with respect to different identity labels. We use man, woman as a concrete example of an identity label set (although this set need not be binary), and present rankings for the labels that are most biased towards one identity or the other. We demonstrate how the statistical properties of different association metrics can lead to different rankings of the most "gender biased" labels, and conclude that normalized pointwise mutual information (nPMI) is most useful in practice. Finally, we announce an open-sourced nPMI visualization tool using TensorBoard.
机器学习中偏差的测量通常侧重于相对于基础真值标签的跨身份子组(如男性和女性)的模型性能。然而,这些方法并不能直接测量模型可能已经学习到的关联,例如标签和身份子组之间的关联。此外,测量模型的偏差需要一个完全注释的评估数据集,这在实践中可能不容易获得。我们提出了一个优雅的数学解决方案,同时解决了这两个问题,使用图像分类作为一个工作示例。通过将分类模型对给定图像的预测视为一组类似于“单词袋”的标签,我们将模型所学到的偏差与不同的身份标签进行排序。我们使用男人,女人作为身份标签集的具体例子(尽管这个集不需要是二元的),并给出最偏向于一个身份或另一个身份的标签的排名。我们展示了不同关联度量的统计属性如何导致最“性别偏见”标签的不同排名,并得出归一化的点互信息(nPMI)在实践中最有用的结论。最后,我们宣布了一个使用TensorBoard的开源nPMI可视化工具。
{"title":"Measuring Model Biases in the Absence of Ground Truth","authors":"Osman Aka, Ken Burke, Alex Bauerle, Christina Greer, Margaret Mitchell","doi":"10.1145/3461702.3462557","DOIUrl":"https://doi.org/10.1145/3461702.3462557","url":null,"abstract":"The measurement of bias in machine learning often focuses on model performance across identity subgroups (such as man and woman) with respect to groundtruth labels. However, these methods do not directly measure the associations that a model may have learned, for example between labels and identity subgroups. Further, measuring a model's bias requires a fully annotated evaluation dataset which may not be easily available in practice. We present an elegant mathematical solution that tackles both issues simultaneously, using image classification as a working example. By treating a classification model's predictions for a given image as a set of labels analogous to a \"bag of words\", we rank the biases that a model has learned with respect to different identity labels. We use man, woman as a concrete example of an identity label set (although this set need not be binary), and present rankings for the labels that are most biased towards one identity or the other. We demonstrate how the statistical properties of different association metrics can lead to different rankings of the most \"gender biased\" labels, and conclude that normalized pointwise mutual information (nPMI) is most useful in practice. Finally, we announce an open-sourced nPMI visualization tool using TensorBoard.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"358 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120966804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Towards Unbiased and Accurate Deferral to Multiple Experts 向多专家提供公正和准确的延期
Pub Date : 2021-02-25 DOI: 10.1145/3461702.3462516
Vijay Keswani, Matthew Lease, K. Kenthapadi
Machine learning models are often implemented in cohort with humans in the pipeline, with the model having an option to defer to a domain expert in cases where it has low confidence in its inference. Our goal is to design mechanisms for ensuring accuracy and fairness in such prediction systems that combine machine learning model inferences and domain expert predictions. Prior work on "deferral systems" in classification settings has focused on the setting of a pipeline with a single expert and aimed to accommodate the inaccuracies and biases of this expert to simultaneously learn an inference model and a deferral system. Our work extends this framework to settings where multiple experts are available, with each expert having their own domain of expertise and biases. We propose a framework that simultaneously learns a classifier and a deferral system, with the deferral system choosing to defer to one or more human experts in cases of input where the classifier has low confidence. We test our framework on a synthetic dataset and a content moderation dataset with biased synthetic experts, and show that it significantly improves the accuracy and fairness of the final predictions, compared to the baselines. We also collect crowdsourced labels for the content moderation task to construct a real-world dataset for the evaluation of hybrid machine-human frameworks and show that our proposed framework outperforms baselines on this real-world dataset as well.
机器学习模型通常是在流水线中与人类一起实现的,在对其推理缺乏信心的情况下,模型可以选择服从领域专家。我们的目标是设计一种机制,以确保这种结合机器学习模型推断和领域专家预测的预测系统的准确性和公平性。先前关于分类设置中的“延迟系统”的工作主要集中在单个专家的管道设置上,旨在适应该专家的不准确性和偏差,以同时学习推理模型和延迟系统。我们的工作将这一框架扩展到多个专家可用的环境中,每个专家都有自己的专业领域和偏见。我们提出了一个同时学习分类器和延迟系统的框架,在分类器置信度低的情况下,延迟系统选择服从一个或多个人类专家的输入。我们在一个合成数据集和一个有偏见的合成专家的内容审核数据集上测试了我们的框架,并表明与基线相比,它显着提高了最终预测的准确性和公平性。我们还为内容审核任务收集众包标签,以构建一个用于评估混合机器-人框架的真实数据集,并表明我们提出的框架在这个真实数据集上也优于基线。
{"title":"Towards Unbiased and Accurate Deferral to Multiple Experts","authors":"Vijay Keswani, Matthew Lease, K. Kenthapadi","doi":"10.1145/3461702.3462516","DOIUrl":"https://doi.org/10.1145/3461702.3462516","url":null,"abstract":"Machine learning models are often implemented in cohort with humans in the pipeline, with the model having an option to defer to a domain expert in cases where it has low confidence in its inference. Our goal is to design mechanisms for ensuring accuracy and fairness in such prediction systems that combine machine learning model inferences and domain expert predictions. Prior work on \"deferral systems\" in classification settings has focused on the setting of a pipeline with a single expert and aimed to accommodate the inaccuracies and biases of this expert to simultaneously learn an inference model and a deferral system. Our work extends this framework to settings where multiple experts are available, with each expert having their own domain of expertise and biases. We propose a framework that simultaneously learns a classifier and a deferral system, with the deferral system choosing to defer to one or more human experts in cases of input where the classifier has low confidence. We test our framework on a synthetic dataset and a content moderation dataset with biased synthetic experts, and show that it significantly improves the accuracy and fairness of the final predictions, compared to the baselines. We also collect crowdsourced labels for the content moderation task to construct a real-world dataset for the evaluation of hybrid machine-human frameworks and show that our proposed framework outperforms baselines on this real-world dataset as well.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125747985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
The Deepfake Detection Dilemma: A Multistakeholder Exploration of Adversarial Dynamics in Synthetic Media 深度伪造检测困境:合成媒体中对抗动态的多利益相关者探索
Pub Date : 2021-02-11 DOI: 10.1145/3461702.3462584
Claire Leibowicz, Sean McGregor, Aviv Ovadya
Synthetic media detection technologies label media as either synthetic or non-synthetic and are increasingly used by journalists, web platforms, and the general public to identify misinformation and other forms of problematic content. As both well-resourced organizations and the non-technical general public generate more sophisticated synthetic media, the capacity for purveyors of problematic content to adapt induces a detection dilemma : as detection practices become more accessible, they become more easily circumvented. This paper describes how a multistakeholder cohort from academia, technology platforms, media entities, and civil society organizations active in synthetic media detection and its socio-technical implications evaluates the detection dilemma. Specifically, we offer an assessment of detection contexts and adversary capacities sourced from the broader, global AI and media integrity community concerned with mitigating the spread of harmful synthetic media. A collection of personas illustrates the intersection between unsophisticated and highly-resourced sponsors of misinformation in the context of their technical capacities. This work concludes that there is no "best'' approach to navigating the detector dilemma, but derives a set of implications from multistakeholder input to better inform detection process decisions and policies, in practice.
合成媒体检测技术将媒体标记为合成或非合成,越来越多地被记者、网络平台和公众用于识别错误信息和其他形式的问题内容。由于资源丰富的组织和非技术的普通公众都产生了更复杂的合成媒体,问题内容的提供者适应的能力导致了检测困境:随着检测实践变得更容易获得,它们变得更容易被绕过。本文描述了来自学术界、技术平台、媒体实体和活跃于合成媒体检测及其社会技术影响的民间社会组织的多方利益相关者群体如何评估检测困境。具体而言,我们对来自更广泛的全球人工智能和媒体完整性社区的检测环境和对手能力进行了评估,这些社区与减轻有害合成媒体的传播有关。一组人物角色说明了在他们的技术能力背景下,不成熟和资源丰富的错误信息赞助者之间的交集。这项工作的结论是,没有“最佳”方法来解决探测器困境,但从多利益相关者的输入中得出了一组含义,以便在实践中更好地为检测过程决策和政策提供信息。
{"title":"The Deepfake Detection Dilemma: A Multistakeholder Exploration of Adversarial Dynamics in Synthetic Media","authors":"Claire Leibowicz, Sean McGregor, Aviv Ovadya","doi":"10.1145/3461702.3462584","DOIUrl":"https://doi.org/10.1145/3461702.3462584","url":null,"abstract":"Synthetic media detection technologies label media as either synthetic or non-synthetic and are increasingly used by journalists, web platforms, and the general public to identify misinformation and other forms of problematic content. As both well-resourced organizations and the non-technical general public generate more sophisticated synthetic media, the capacity for purveyors of problematic content to adapt induces a detection dilemma : as detection practices become more accessible, they become more easily circumvented. This paper describes how a multistakeholder cohort from academia, technology platforms, media entities, and civil society organizations active in synthetic media detection and its socio-technical implications evaluates the detection dilemma. Specifically, we offer an assessment of detection contexts and adversary capacities sourced from the broader, global AI and media integrity community concerned with mitigating the spread of harmful synthetic media. A collection of personas illustrates the intersection between unsophisticated and highly-resourced sponsors of misinformation in the context of their technical capacities. This work concludes that there is no \"best'' approach to navigating the detector dilemma, but derives a set of implications from multistakeholder input to better inform detection process decisions and policies, in practice.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"258 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127543364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Learning to Generate Fair Clusters from Demonstrations 学习从演示中生成公平的集群
Pub Date : 2021-02-08 DOI: 10.1145/3461702.3462558
Sainyam Galhotra, Sandhya Saisubramanian, S. Zilberstein
Fair clustering is the process of grouping similar entities together, while satisfying a mathematically well-defined fairness metric as a constraint. Due to the practical challenges in precise model specification, the prescribed fairness constraints are often incomplete and act as proxies to the intended fairness requirement. Clustering with proxies may lead to biased outcomes when the system is deployed. We examine how to identify the intended fairness constraint for a problem based on limited demonstrations from an expert. Each demonstration is a clustering over a subset of the data. We present an algorithm to identify the fairness metric from demonstrations and generate clusters using existing off-the-shelf clustering techniques, and analyze its theoretical properties. To extend our approach to novel fairness metrics for which clustering algorithms do not currently exist, we present a greedy method for clustering. Additionally, we investigate how to generate interpretable solutions using our approach. Empirical evaluation on three real-world datasets demonstrates the effectiveness of our approach in quickly identifying the underlying fairness and interpretability constraints, which are then used to generate fair and interpretable clusters.
公平聚类是将相似实体分组在一起的过程,同时满足数学上定义良好的公平度量作为约束。由于精确模型规范中的实际挑战,规定的公平性约束通常是不完整的,并且充当预期公平性需求的代理。在部署系统时,使用代理进行集群可能会导致有偏差的结果。我们研究如何根据专家的有限演示来确定问题的预期公平性约束。每个演示都是对数据子集的聚类。我们提出了一种从演示中识别公平度量的算法,并使用现有的现成聚类技术生成聚类,并分析了其理论性质。为了将我们的方法扩展到目前尚不存在聚类算法的新型公平度量,我们提出了一种贪婪聚类方法。此外,我们还研究了如何使用我们的方法生成可解释的解决方案。对三个真实世界数据集的实证评估表明,我们的方法在快速识别潜在的公平性和可解释性约束方面是有效的,然后将其用于生成公平和可解释的聚类。
{"title":"Learning to Generate Fair Clusters from Demonstrations","authors":"Sainyam Galhotra, Sandhya Saisubramanian, S. Zilberstein","doi":"10.1145/3461702.3462558","DOIUrl":"https://doi.org/10.1145/3461702.3462558","url":null,"abstract":"Fair clustering is the process of grouping similar entities together, while satisfying a mathematically well-defined fairness metric as a constraint. Due to the practical challenges in precise model specification, the prescribed fairness constraints are often incomplete and act as proxies to the intended fairness requirement. Clustering with proxies may lead to biased outcomes when the system is deployed. We examine how to identify the intended fairness constraint for a problem based on limited demonstrations from an expert. Each demonstration is a clustering over a subset of the data. We present an algorithm to identify the fairness metric from demonstrations and generate clusters using existing off-the-shelf clustering techniques, and analyze its theoretical properties. To extend our approach to novel fairness metrics for which clustering algorithms do not currently exist, we present a greedy method for clustering. Additionally, we investigate how to generate interpretable solutions using our approach. Empirical evaluation on three real-world datasets demonstrates the effectiveness of our approach in quickly identifying the underlying fairness and interpretability constraints, which are then used to generate fair and interpretable clusters.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127045640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Fairness for Unobserved Characteristics: Insights from Technological Impacts on Queer Communities 未观察特征的公平性:科技对酷儿群体的影响
Pub Date : 2021-02-03 DOI: 10.1145/3461702.3462540
Nenad Tomašev, Kevin R. McKee, J. Kay, Shakir Mohamed
Advances in algorithmic fairness have largely omitted sexual orientation and gender identity. We explore queer concerns in privacy, censorship, language, online safety, health, and employment to study the positive and negative effects of artificial intelligence on queer communities. These issues underscore the need for new directions in fairness research that take into account a multiplicity of considerations, from privacy preservation, context sensitivity and process fairness, to an awareness of sociotechnical impact and the increasingly important role of inclusive and participatory research processes. Most current approaches for algorithmic fairness assume that the target characteristics for fairness---frequently, race and legal gender---can be observed or recorded. Sexual orientation and gender identity are prototypical instances of unobserved characteristics, which are frequently missing, unknown or fundamentally unmeasurable. This paper highlights the importance of developing new approaches for algorithmic fairness that break away from the prevailing assumption of observed characteristics.
算法公平性的进步在很大程度上忽略了性取向和性别认同。我们探讨酷儿在隐私、审查、语言、网络安全、健康和就业方面的关注,研究人工智能对酷儿社区的积极和消极影响。这些问题强调了公平研究需要新的方向,需要考虑到多种因素,从隐私保护、上下文敏感性和过程公平性,到对社会技术影响的认识以及包容性和参与性研究过程日益重要的作用。目前大多数算法公平的方法都假设公平的目标特征——通常是种族和法律性别——可以被观察或记录。性取向和性别认同是未被观察到的特征的典型实例,这些特征往往是缺失的、未知的或根本无法测量的。本文强调了开发算法公平性新方法的重要性,这些方法打破了对观察特征的普遍假设。
{"title":"Fairness for Unobserved Characteristics: Insights from Technological Impacts on Queer Communities","authors":"Nenad Tomašev, Kevin R. McKee, J. Kay, Shakir Mohamed","doi":"10.1145/3461702.3462540","DOIUrl":"https://doi.org/10.1145/3461702.3462540","url":null,"abstract":"Advances in algorithmic fairness have largely omitted sexual orientation and gender identity. We explore queer concerns in privacy, censorship, language, online safety, health, and employment to study the positive and negative effects of artificial intelligence on queer communities. These issues underscore the need for new directions in fairness research that take into account a multiplicity of considerations, from privacy preservation, context sensitivity and process fairness, to an awareness of sociotechnical impact and the increasingly important role of inclusive and participatory research processes. Most current approaches for algorithmic fairness assume that the target characteristics for fairness---frequently, race and legal gender---can be observed or recorded. Sexual orientation and gender identity are prototypical instances of unobserved characteristics, which are frequently missing, unknown or fundamentally unmeasurable. This paper highlights the importance of developing new approaches for algorithmic fairness that break away from the prevailing assumption of observed characteristics.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"121 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130116897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 56
Quantum Fair Machine Learning 量子公平机器学习
Pub Date : 2021-02-01 DOI: 10.1145/3461702.3462611
Elija Perrier
In this paper, we inaugurate the field of quantum fair machine learning. We undertake a comparative analysis of differences and similarities between classical and quantum fair machine learning algorithms, specifying how the unique features of quantum computation alter measures, metrics and remediation strategies when quantum algorithms are subject to fairness constraints. We present the first results in quantum fair machine learning by demonstrating the use of Grover's search algorithm to satisfy statistical parity constraints imposed on quantum algorithms. We provide lower-bounds on iterations needed to achieve such statistical parity within ε-tolerance. We extend canonical Lipschitz-conditioned individual fairness criteria to the quantum setting using quantum metrics. We examine the consequences for typical measures of fairness in machine learning context when quantum information processing and quantum data are involved. Finally, we propose open questions and research programmes for this new field of interest to researchers in computer science, ethics and quantum computation.
在本文中,我们开创了量子公平机器学习领域。我们对经典和量子公平机器学习算法之间的异同进行了比较分析,说明了当量子算法受到公平约束时,量子计算的独特特征如何改变度量、指标和补救策略。我们通过演示使用Grover搜索算法来满足施加在量子算法上的统计奇偶性约束,提出了量子公平机器学习的第一个结果。我们提供了在ε-容差范围内实现这种统计奇偶性所需的迭代的下界。我们使用量子度量将标准李普希茨条件下的个体公平性标准扩展到量子设置。当涉及量子信息处理和量子数据时,我们研究了机器学习环境中典型的公平度量的后果。最后,我们向计算机科学、伦理学和量子计算领域的研究人员提出了这个新领域的开放问题和研究计划。
{"title":"Quantum Fair Machine Learning","authors":"Elija Perrier","doi":"10.1145/3461702.3462611","DOIUrl":"https://doi.org/10.1145/3461702.3462611","url":null,"abstract":"In this paper, we inaugurate the field of quantum fair machine learning. We undertake a comparative analysis of differences and similarities between classical and quantum fair machine learning algorithms, specifying how the unique features of quantum computation alter measures, metrics and remediation strategies when quantum algorithms are subject to fairness constraints. We present the first results in quantum fair machine learning by demonstrating the use of Grover's search algorithm to satisfy statistical parity constraints imposed on quantum algorithms. We provide lower-bounds on iterations needed to achieve such statistical parity within ε-tolerance. We extend canonical Lipschitz-conditioned individual fairness criteria to the quantum setting using quantum metrics. We examine the consequences for typical measures of fairness in machine learning context when quantum information processing and quantum data are involved. Finally, we propose open questions and research programmes for this new field of interest to researchers in computer science, ethics and quantum computation.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129068205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Emergent Unfairness in Algorithmic Fairness-Accuracy Trade-Off Research 算法公平-精度权衡研究中的突现不公平
Pub Date : 2021-02-01 DOI: 10.1145/3461702.3462519
A. Feder Cooper, Ellen Abrams, NA Na
Across machine learning (ML) sub-disciplines, researchers make explicit mathematical assumptions in order to facilitate proof-writing. We note that, specifically in the area of fairness-accuracy trade-off optimization scholarship, similar attention is not paid to the normative assumptions that ground this approach. Such assumptions presume that 1) accuracy and fairness are in inherent opposition to one another, 2) strict notions of mathematical equality can adequately model fairness, 3) it is possible to measure the accuracy and fairness of decisions independent from historical context, and 4) collecting more data on marginalized individuals is a reasonable solution to mitigate the effects of the trade-off. We argue that such assumptions, which are often left implicit and unexamined, lead to inconsistent conclusions: While the intended goal of this work may be to improve the fairness of machine learning models, these unexamined, implicit assumptions can in fact result in emergent unfairness. We conclude by suggesting a concrete path forward toward a potential resolution.
在机器学习(ML)子学科中,研究人员做出明确的数学假设,以促进证明写作。我们注意到,特别是在公平-准确性权衡优化学术领域,没有对这种方法的规范性假设给予类似的关注。这些假设假设1)准确性和公平性是内在对立的,2)严格的数学平等概念可以充分地模拟公平性,3)独立于历史背景来衡量决策的准确性和公平性是可能的,4)收集更多边缘化个体的数据是减轻权衡影响的合理解决方案。我们认为,这些假设通常是隐含的和未经检验的,导致结论不一致:虽然这项工作的预期目标可能是提高机器学习模型的公平性,但这些未经检验的隐含假设实际上可能导致紧急不公平。最后,我们提出了一个可能解决问题的具体途径。
{"title":"Emergent Unfairness in Algorithmic Fairness-Accuracy Trade-Off Research","authors":"A. Feder Cooper, Ellen Abrams, NA Na","doi":"10.1145/3461702.3462519","DOIUrl":"https://doi.org/10.1145/3461702.3462519","url":null,"abstract":"Across machine learning (ML) sub-disciplines, researchers make explicit mathematical assumptions in order to facilitate proof-writing. We note that, specifically in the area of fairness-accuracy trade-off optimization scholarship, similar attention is not paid to the normative assumptions that ground this approach. Such assumptions presume that 1) accuracy and fairness are in inherent opposition to one another, 2) strict notions of mathematical equality can adequately model fairness, 3) it is possible to measure the accuracy and fairness of decisions independent from historical context, and 4) collecting more data on marginalized individuals is a reasonable solution to mitigate the effects of the trade-off. We argue that such assumptions, which are often left implicit and unexamined, lead to inconsistent conclusions: While the intended goal of this work may be to improve the fairness of machine learning models, these unexamined, implicit assumptions can in fact result in emergent unfairness. We conclude by suggesting a concrete path forward toward a potential resolution.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"222 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131508401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
RAWLSNET: Altering Bayesian Networks to Encode Rawlsian Fair Equality of Opportunity RAWLSNET:改变贝叶斯网络编码罗尔斯公平机会均等
Pub Date : 2021-01-31 DOI: 10.1145/3461702.3462618
David Liu, Zohair Shafi, W. Fleisher, Tina Eliassi-Rad, Scott Alfeld
We present RAWLSNET, a system for altering Bayesian Network (BN) models to satisfy the Rawlsian principle of fair equality of opportunity (FEO). RAWLSNET's BN models generate aspirational data distributions: data generated to reflect an ideally fair, FEO-satisfying society. FEO states that everyone with the same talent and willingness to use it should have the same chance of achieving advantageous social positions (e.g., employment), regardless of their background circumstances (e.g., socioeconomic status). Satisfying FEO requires alterations to social structures such as school assignments. Our paper describes RAWLSNET, a method which takes as input a BN representation of an FEO application and alters the BN's parameters so as to satisfy FEO when possible, and minimize deviation from FEO otherwise. We also offer guidance for applying RAWLSNET, including on recognizing proper applications of FEO. We demonstrate the use of RAWLSNET with publicly available data sets. RAWLSNET's altered BNs offer the novel capability of generating aspirational data for FEO-relevant tasks. Aspirational data are free from biases of real-world data, and thus are useful for recognizing and detecting sources of unfairness in machine learning algorithms besides biased data.
RAWLSNET是一个改进贝叶斯网络(BN)模型以满足罗尔斯公平机会均等原则的系统。RAWLSNET的BN模型生成理想的数据分布:生成的数据反映理想的公平,feo满意的社会。FEO指出,每个人都有同样的才能和意愿去使用它,应该有同样的机会获得有利的社会地位(例如,就业),而不管他们的背景环境(例如,社会经济地位)。满足FEO需要改变社会结构,比如学校作业。本文描述了RAWLSNET方法,该方法将FEO应用的BN表示作为输入,并改变BN的参数,使其尽可能满足FEO,否则尽量减少与FEO的偏差。我们还提供了应用RAWLSNET的指导,包括识别FEO的适当应用。我们演示了使用公开可用的数据集来使用RAWLSNET。RAWLSNET的改进型bn提供了为feo相关任务生成理想数据的新能力。理想数据没有现实世界数据的偏见,因此除了偏见数据之外,对于识别和检测机器学习算法中的不公平来源非常有用。
{"title":"RAWLSNET: Altering Bayesian Networks to Encode Rawlsian Fair Equality of Opportunity","authors":"David Liu, Zohair Shafi, W. Fleisher, Tina Eliassi-Rad, Scott Alfeld","doi":"10.1145/3461702.3462618","DOIUrl":"https://doi.org/10.1145/3461702.3462618","url":null,"abstract":"We present RAWLSNET, a system for altering Bayesian Network (BN) models to satisfy the Rawlsian principle of fair equality of opportunity (FEO). RAWLSNET's BN models generate aspirational data distributions: data generated to reflect an ideally fair, FEO-satisfying society. FEO states that everyone with the same talent and willingness to use it should have the same chance of achieving advantageous social positions (e.g., employment), regardless of their background circumstances (e.g., socioeconomic status). Satisfying FEO requires alterations to social structures such as school assignments. Our paper describes RAWLSNET, a method which takes as input a BN representation of an FEO application and alters the BN's parameters so as to satisfy FEO when possible, and minimize deviation from FEO otherwise. We also offer guidance for applying RAWLSNET, including on recognizing proper applications of FEO. We demonstrate the use of RAWLSNET with publicly available data sets. RAWLSNET's altered BNs offer the novel capability of generating aspirational data for FEO-relevant tasks. Aspirational data are free from biases of real-world data, and thus are useful for recognizing and detecting sources of unfairness in machine learning algorithms besides biased data.","PeriodicalId":197336,"journal":{"name":"Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130286252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1