首页 > 最新文献

Proceedings of the Conference on Fairness, Accountability, and Transparency最新文献

英文 中文
SIREN: A Simulation Framework for Understanding the Effects of Recommender Systems in Online News Environments SIREN:用于理解在线新闻环境中推荐系统影响的模拟框架
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287583
D. Bountouridis, Jaron Harambam, M. Makhortykh, M. Marrero, N. Tintarev, C. Hauff
The growing volume of digital data stimulates the adoption of recommender systems in different socioeconomic domains, including news industries. While news recommenders help consumers deal with information overload and increase their engagement, their use also raises an increasing number of societal concerns, such as "Matthew effects", "filter bubbles", and the overall lack of transparency. We argue that focusing on transparency for content-providers is an under-explored avenue. As such, we designed a simulation framework called SIREN1 (SImulating Recommender Effects in online News environments), that allows content providers to (i) select and parameterize different recommenders and (ii) analyze and visualize their effects with respect to two diversity metrics. Taking the U.S. news media as a case study, we present an analysis on the recommender effects with respect to long-tail novelty and unexpectedness using SIREN. Our analysis offers a number of interesting findings, such as the similar potential of certain algorithmically simple (item-based k-Nearest Neighbour) and sophisticated strategies (based on Bayesian Personalized Ranking) to increase diversity over time. Overall, we argue that simulating the effects of recommender systems can help content providers to make more informed decisions when choosing algorithmic recommenders, and as such can help mitigate the aforementioned societal concerns.
不断增长的数字数据量刺激了不同社会经济领域(包括新闻行业)对推荐系统的采用。新闻推荐在帮助消费者处理信息过载、提高参与度的同时,也引发了越来越多的社会担忧,比如“马太效应”、“过滤泡沫”,以及整体缺乏透明度。我们认为,关注内容提供商的透明度是一个尚未得到充分探索的途径。因此,我们设计了一个名为SIREN1(模拟在线新闻环境中的推荐效果)的模拟框架,它允许内容提供商(i)选择和参数化不同的推荐者,(ii)根据两个多样性指标分析和可视化他们的效果。本文以美国新闻媒体为例,利用SIREN分析了长尾新颖性和意外性方面的推荐效应。我们的分析提供了许多有趣的发现,例如某些简单算法(基于物品的k近邻)和复杂策略(基于贝叶斯个性化排名)在增加多样性方面具有相似的潜力。总的来说,我们认为模拟推荐系统的效果可以帮助内容提供商在选择算法推荐时做出更明智的决定,因此可以帮助减轻上述社会问题。
{"title":"SIREN: A Simulation Framework for Understanding the Effects of Recommender Systems in Online News Environments","authors":"D. Bountouridis, Jaron Harambam, M. Makhortykh, M. Marrero, N. Tintarev, C. Hauff","doi":"10.1145/3287560.3287583","DOIUrl":"https://doi.org/10.1145/3287560.3287583","url":null,"abstract":"The growing volume of digital data stimulates the adoption of recommender systems in different socioeconomic domains, including news industries. While news recommenders help consumers deal with information overload and increase their engagement, their use also raises an increasing number of societal concerns, such as \"Matthew effects\", \"filter bubbles\", and the overall lack of transparency. We argue that focusing on transparency for content-providers is an under-explored avenue. As such, we designed a simulation framework called SIREN1 (SImulating Recommender Effects in online News environments), that allows content providers to (i) select and parameterize different recommenders and (ii) analyze and visualize their effects with respect to two diversity metrics. Taking the U.S. news media as a case study, we present an analysis on the recommender effects with respect to long-tail novelty and unexpectedness using SIREN. Our analysis offers a number of interesting findings, such as the similar potential of certain algorithmically simple (item-based k-Nearest Neighbour) and sophisticated strategies (based on Bayesian Personalized Ranking) to increase diversity over time. Overall, we argue that simulating the effects of recommender systems can help content providers to make more informed decisions when choosing algorithmic recommenders, and as such can help mitigate the aforementioned societal concerns.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89699350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
Controlling Polarization in Personalization: An Algorithmic Framework 控制个性化中的极化:一个算法框架
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287601
L. E. Celis, Sayash Kapoor, Farnood Salehi, Nisheeth K. Vishnoi
Personalization is pervasive in the online space as it leads to higher efficiency for the user and higher revenue for the platform by individualizing the most relevant content for each user. However, recent studies suggest that such personalization can learn and propagate systemic biases and polarize opinions; this has led to calls for regulatory mechanisms and algorithms that are constrained to combat bias and the resulting echo-chamber effect. We propose a versatile framework that allows for the possibility to reduce polarization in personalized systems by allowing the user to constrain the distribution from which content is selected. We then present a scalable algorithm with provable guarantees that satisfies the given constraints on the types of the content that can be displayed to a user, but -- subject to these constraints -- will continue to learn and personalize the content in order to maximize utility. We illustrate this framework on a curated dataset of online news articles that are conservative or liberal, show that it can control polarization, and examine the trade-off between decreasing polarization and the resulting loss to revenue. We further exhibit the flexibility and scalability of our approach by framing the problem in terms of the more general diverse content selection problem and test it empirically on both a News dataset and the MovieLens dataset.
个性化在在线空间中非常普遍,因为它通过为每个用户个性化最相关的内容,为用户带来更高的效率,并为平台带来更高的收入。然而,最近的研究表明,这种个性化可以学习和传播系统性偏见和两极分化的意见;这导致了对监管机制和算法的呼吁,这些机制和算法受到限制,以对抗偏见和由此产生的回声室效应。我们提出了一个通用的框架,通过允许用户约束选择内容的分布,可以减少个性化系统中的两极分化。然后,我们提出了一种可扩展的算法,具有可证明的保证,满足可以显示给用户的内容类型的给定约束,但是——受这些约束的约束——将继续学习和个性化内容,以最大化效用。我们在保守或自由的在线新闻文章的策划数据集上说明了这个框架,表明它可以控制两极分化,并检查两极分化减少与由此导致的收入损失之间的权衡。我们进一步展示了我们方法的灵活性和可扩展性,方法是根据更一般的多样化内容选择问题来构建问题,并在News数据集和MovieLens数据集上进行经验测试。
{"title":"Controlling Polarization in Personalization: An Algorithmic Framework","authors":"L. E. Celis, Sayash Kapoor, Farnood Salehi, Nisheeth K. Vishnoi","doi":"10.1145/3287560.3287601","DOIUrl":"https://doi.org/10.1145/3287560.3287601","url":null,"abstract":"Personalization is pervasive in the online space as it leads to higher efficiency for the user and higher revenue for the platform by individualizing the most relevant content for each user. However, recent studies suggest that such personalization can learn and propagate systemic biases and polarize opinions; this has led to calls for regulatory mechanisms and algorithms that are constrained to combat bias and the resulting echo-chamber effect. We propose a versatile framework that allows for the possibility to reduce polarization in personalized systems by allowing the user to constrain the distribution from which content is selected. We then present a scalable algorithm with provable guarantees that satisfies the given constraints on the types of the content that can be displayed to a user, but -- subject to these constraints -- will continue to learn and personalize the content in order to maximize utility. We illustrate this framework on a curated dataset of online news articles that are conservative or liberal, show that it can control polarization, and examine the trade-off between decreasing polarization and the resulting loss to revenue. We further exhibit the flexibility and scalability of our approach by framing the problem in terms of the more general diverse content selection problem and test it empirically on both a News dataset and the MovieLens dataset.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83245245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
Fairness and Abstraction in Sociotechnical Systems 社会技术系统中的公平与抽象
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287598
Andrew D. Selbst, D. Boyd, Sorelle A. Friedler, Suresh Venkatasubramanian, J. Vertesi
A key goal of the fair-ML community is to develop machine-learning based systems that, once introduced into a social context, can achieve social and legal outcomes such as fairness, justice, and due process. Bedrock concepts in computer science---such as abstraction and modular design---are used to define notions of fairness and discrimination, to produce fairness-aware learning algorithms, and to intervene at different stages of a decision-making pipeline to produce "fair" outcomes. In this paper, however, we contend that these concepts render technical interventions ineffective, inaccurate, and sometimes dangerously misguided when they enter the societal context that surrounds decision-making systems. We outline this mismatch with five "traps" that fair-ML work can fall into even as it attempts to be more context-aware in comparison to traditional data science. We draw on studies of sociotechnical systems in Science and Technology Studies to explain why such traps occur and how to avoid them. Finally, we suggest ways in which technical designers can mitigate the traps through a refocusing of design in terms of process rather than solutions, and by drawing abstraction boundaries to include social actors rather than purely technical ones.
公平机器学习社区的一个关键目标是开发基于机器学习的系统,一旦引入社会环境,就可以实现社会和法律结果,如公平、正义和正当程序。计算机科学中的基本概念——如抽象和模块化设计——被用来定义公平和歧视的概念,产生公平意识的学习算法,并在决策管道的不同阶段进行干预,以产生“公平”的结果。然而,在本文中,我们认为,这些概念使得技术干预无效,不准确,有时当它们进入围绕决策系统的社会背景时,会产生危险的误导。我们用五个“陷阱”概述了这种不匹配,即使与传统数据科学相比,公平机器学习工作也可能陷入这些“陷阱”。我们利用科学与技术研究中的社会技术系统研究来解释为什么会发生这种陷阱以及如何避免它们。最后,我们建议技术设计师可以通过从过程而不是解决方案的角度重新关注设计,并通过绘制抽象边界来包括社会参与者而不是纯粹的技术参与者来减轻陷阱。
{"title":"Fairness and Abstraction in Sociotechnical Systems","authors":"Andrew D. Selbst, D. Boyd, Sorelle A. Friedler, Suresh Venkatasubramanian, J. Vertesi","doi":"10.1145/3287560.3287598","DOIUrl":"https://doi.org/10.1145/3287560.3287598","url":null,"abstract":"A key goal of the fair-ML community is to develop machine-learning based systems that, once introduced into a social context, can achieve social and legal outcomes such as fairness, justice, and due process. Bedrock concepts in computer science---such as abstraction and modular design---are used to define notions of fairness and discrimination, to produce fairness-aware learning algorithms, and to intervene at different stages of a decision-making pipeline to produce \"fair\" outcomes. In this paper, however, we contend that these concepts render technical interventions ineffective, inaccurate, and sometimes dangerously misguided when they enter the societal context that surrounds decision-making systems. We outline this mismatch with five \"traps\" that fair-ML work can fall into even as it attempts to be more context-aware in comparison to traditional data science. We draw on studies of sociotechnical systems in Science and Technology Studies to explain why such traps occur and how to avoid them. Finally, we suggest ways in which technical designers can mitigate the traps through a refocusing of design in terms of process rather than solutions, and by drawing abstraction boundaries to include social actors rather than purely technical ones.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77045777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 641
The Profiling Potential of Computer Vision and the Challenge of Computational Empiricism 计算机视觉的轮廓分析潜力和计算经验主义的挑战
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287568
Jake Goldenfein
Computer vision and other biometrics data science applications have commenced a new project of profiling people. Rather than using 'transaction generated information', these systems measure the 'real world' and produce an assessment of the 'world state' - in this case an assessment of some individual trait. Instead of using proxies or scores to evaluate people, they increasingly deploy a logic of revealing the truth about reality and the people within it. While these profiling knowledge claims are sometimes tentative, they increasingly suggest that only through computation can these excesses of reality be captured and understood. This article explores the bases of those claims in the systems of measurement, representation, and classification deployed in computer vision. It asks if there is something new in this type of knowledge claim, sketches an account of a new form of computational empiricism being operationalised, and questions what kind of human subject is being constructed by these technological systems and practices. Finally, the article explores legal mechanisms for contesting the emergence of computational empiricism as the dominant knowledge platform for understanding the world and the people within it.
计算机视觉和其他生物识别数据科学应用已经开始了一个分析人的新项目。这些系统不是使用“交易生成的信息”,而是测量“真实世界”并产生对“世界状态”的评估——在这种情况下是对某些个体特征的评估。他们不再使用代理或分数来评估人,而是越来越多地采用一种揭示现实真相和现实中的人的逻辑。虽然这些描述知识的主张有时是试探性的,但它们越来越多地表明,只有通过计算才能捕捉和理解这些过度的现实。本文探讨了在计算机视觉中部署的测量、表示和分类系统中这些主张的基础。它询问这种类型的知识主张中是否有新的东西,描绘了一种新的计算经验主义的运作形式,并质疑这些技术系统和实践正在构建什么样的人类主体。最后,本文探讨了法律机制,以对抗计算经验主义作为理解世界及其中的人的主要知识平台的出现。
{"title":"The Profiling Potential of Computer Vision and the Challenge of Computational Empiricism","authors":"Jake Goldenfein","doi":"10.1145/3287560.3287568","DOIUrl":"https://doi.org/10.1145/3287560.3287568","url":null,"abstract":"Computer vision and other biometrics data science applications have commenced a new project of profiling people. Rather than using 'transaction generated information', these systems measure the 'real world' and produce an assessment of the 'world state' - in this case an assessment of some individual trait. Instead of using proxies or scores to evaluate people, they increasingly deploy a logic of revealing the truth about reality and the people within it. While these profiling knowledge claims are sometimes tentative, they increasingly suggest that only through computation can these excesses of reality be captured and understood. This article explores the bases of those claims in the systems of measurement, representation, and classification deployed in computer vision. It asks if there is something new in this type of knowledge claim, sketches an account of a new form of computational empiricism being operationalised, and questions what kind of human subject is being constructed by these technological systems and practices. Finally, the article explores legal mechanisms for contesting the emergence of computational empiricism as the dominant knowledge platform for understanding the world and the people within it.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80220759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Beyond Open vs. Closed: Balancing Individual Privacy and Public Accountability in Data Sharing 超越开放与封闭:在数据共享中平衡个人隐私和公共责任
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287577
Meg Young, Luke Rodriguez, Emilyann Keller, Feiyang Sun, Boyang Sa, Jan Whittington, Bill Howe
Data too sensitive to be "open" for analysis and re-purposing typically remains "closed" as proprietary information. This dichotomy undermines efforts to make algorithmic systems more fair, transparent, and accountable. Access to proprietary data in particular is needed by government agencies to enforce policy, researchers to evaluate methods, and the public to hold agencies accountable; all of these needs must be met while preserving individual privacy and firm competitiveness. In this paper, we describe an integrated legal-technical approach provided by a third-party public-private data trust designed to balance these competing interests. Basic membership allows firms and agencies to enable low-risk access to data for compliance reporting and core methods research, while modular data sharing agreements support a wide array of projects and use cases. Unless specifically stated otherwise in an agreement, all data access is initially provided to end users through customized synthetic datasets that offer a) strong privacy guarantees, b) removal of signals that could expose competitive advantage, and c) removal of biases that could reinforce discriminatory policies, all while maintaining fidelity to the original data. We find that using synthetic data in conjunction with strong legal protections over raw data strikes a balance between transparency, proprietorship, privacy, and research objectives. This legal-technical framework can form the basis for data trusts in a variety of contexts.
过于敏感而不能“开放”进行分析和重新利用的数据通常作为专有信息保持“封闭”。这种二分法破坏了使算法系统更加公平、透明和负责任的努力。政府机构需要获取专有数据来执行政策,研究人员需要评估方法,公众需要对机构问责;所有这些需求都必须在保护个人隐私和企业竞争力的同时得到满足。在本文中,我们描述了一种由第三方公私数据信托提供的综合法律技术方法,旨在平衡这些相互竞争的利益。基本会员资格使公司和机构能够以低风险访问数据以进行合规报告和核心方法研究,而模块化数据共享协议支持广泛的项目和用例。除非在协议中另有明确规定,所有数据访问最初都是通过定制的合成数据集提供给最终用户的,这些数据集提供a)强有力的隐私保障,b)去除可能暴露竞争优势的信号,以及c)去除可能加强歧视性政策的偏见,同时保持对原始数据的忠实。我们发现,将合成数据与对原始数据的强有力的法律保护相结合,可以在透明度、所有权、隐私和研究目标之间取得平衡。这种法律-技术框架可以构成各种情况下数据信任的基础。
{"title":"Beyond Open vs. Closed: Balancing Individual Privacy and Public Accountability in Data Sharing","authors":"Meg Young, Luke Rodriguez, Emilyann Keller, Feiyang Sun, Boyang Sa, Jan Whittington, Bill Howe","doi":"10.1145/3287560.3287577","DOIUrl":"https://doi.org/10.1145/3287560.3287577","url":null,"abstract":"Data too sensitive to be \"open\" for analysis and re-purposing typically remains \"closed\" as proprietary information. This dichotomy undermines efforts to make algorithmic systems more fair, transparent, and accountable. Access to proprietary data in particular is needed by government agencies to enforce policy, researchers to evaluate methods, and the public to hold agencies accountable; all of these needs must be met while preserving individual privacy and firm competitiveness. In this paper, we describe an integrated legal-technical approach provided by a third-party public-private data trust designed to balance these competing interests. Basic membership allows firms and agencies to enable low-risk access to data for compliance reporting and core methods research, while modular data sharing agreements support a wide array of projects and use cases. Unless specifically stated otherwise in an agreement, all data access is initially provided to end users through customized synthetic datasets that offer a) strong privacy guarantees, b) removal of signals that could expose competitive advantage, and c) removal of biases that could reinforce discriminatory policies, all while maintaining fidelity to the original data. We find that using synthetic data in conjunction with strong legal protections over raw data strikes a balance between transparency, proprietorship, privacy, and research objectives. This legal-technical framework can form the basis for data trusts in a variety of contexts.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87680901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Who's the Guinea Pig?: Investigating Online A/B/n Tests in-the-Wild 谁是豚鼠?在野外调查在线A/B/n测试
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287565
Shan Jiang, John Martin, Christo Wilson
A/B/n testing has been adopted by many technology companies as a data-driven approach to product design and optimization. These tests are often run on their websites without explicit consent from users. In this paper, we investigate such online A/B/n tests by using Optimizely as a lens. First, we provide measurement results of 575 websites that use Optimizely drawn from the Alexa Top-1M, and analyze the distributions of their audiences and experiments. Then, we use three case studies to discuss potential ethical pitfalls of such experiments, including involvement of political content, price discrimination, and advertising campaigns. We conclude with a suggestion for greater awareness of ethical concerns inherent in human experimentation and a call for increased transparency among A/B/n test operators.
A/B/n测试已被许多科技公司采用,作为产品设计和优化的数据驱动方法。这些测试通常在未经用户明确同意的情况下在他们的网站上进行。在本文中,我们以Optimizely为透镜来研究这种在线A/B/n测试。首先,我们从Alexa Top-1M中提取了575个使用Optimizely的网站的测量结果,并分析了其受众和实验的分布。然后,我们使用三个案例研究来讨论这些实验的潜在伦理陷阱,包括政治内容的参与、价格歧视和广告活动。最后,我们建议提高对人体实验中固有的道德问题的认识,并呼吁增加a /B/n测试操作者的透明度。
{"title":"Who's the Guinea Pig?: Investigating Online A/B/n Tests in-the-Wild","authors":"Shan Jiang, John Martin, Christo Wilson","doi":"10.1145/3287560.3287565","DOIUrl":"https://doi.org/10.1145/3287560.3287565","url":null,"abstract":"A/B/n testing has been adopted by many technology companies as a data-driven approach to product design and optimization. These tests are often run on their websites without explicit consent from users. In this paper, we investigate such online A/B/n tests by using Optimizely as a lens. First, we provide measurement results of 575 websites that use Optimizely drawn from the Alexa Top-1M, and analyze the distributions of their audiences and experiments. Then, we use three case studies to discuss potential ethical pitfalls of such experiments, including involvement of political content, price discrimination, and advertising campaigns. We conclude with a suggestion for greater awareness of ethical concerns inherent in human experimentation and a call for increased transparency among A/B/n test operators.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85562363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Disparate Interactions: An Algorithm-in-the-Loop Analysis of Fairness in Risk Assessments 不同的相互作用:风险评估公平性的循环算法分析
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287563
Ben Green, Yiling Chen
Despite vigorous debates about the technical characteristics of risk assessments being deployed in the U.S. criminal justice system, remarkably little research has studied how these tools affect actual decision-making processes. After all, risk assessments do not make definitive decisions---they inform judges, who are the final arbiters. It is therefore essential that considerations of risk assessments be informed by rigorous studies of how judges actually interpret and use them. This paper takes a first step toward such research on human interactions with risk assessments through a controlled experimental study on Amazon Mechanical Turk. We found several behaviors that call into question the supposed efficacy and fairness of risk assessments: our study participants 1) underperformed the risk assessment even when presented with its predictions, 2) could not effectively evaluate the accuracy of their own or the risk assessment's predictions, and 3) exhibited behaviors fraught with "disparate interactions," whereby the use of risk assessments led to higher risk predictions about black defendants and lower risk predictions about white defendants. These results suggest the need for a new "algorithm-in-the-loop" framework that places machine learning decision-making aids into the sociotechnical context of improving human decisions rather than the technical context of generating the best prediction in the abstract. If risk assessments are to be used at all, they must be grounded in rigorous evaluations of their real-world impacts instead of in their theoretical potential.
尽管对美国刑事司法系统中部署的风险评估的技术特征进行了激烈的辩论,但关于这些工具如何影响实际决策过程的研究却非常少。毕竟,风险评估并不能做出决定性的决定——它们只是告知作为最终仲裁者的法官。因此,必须通过对法官如何实际解释和使用风险评估的严格研究,为风险评估的考虑提供信息。本文通过对Amazon Mechanical Turk的对照实验研究,迈出了人类互动与风险评估研究的第一步。我们发现了一些行为,这些行为对风险评估的有效性和公平性提出了质疑:我们的研究参与者1)即使提供了风险评估的预测,他们的表现也不佳;2)不能有效地评估他们自己或风险评估预测的准确性;3)表现出充满“不同的相互作用”的行为,即使用风险评估导致对黑人被告的风险预测较高,对白人被告的风险预测较低。这些结果表明,需要一个新的“循环算法”框架,将机器学习决策辅助工具置于改善人类决策的社会技术背景中,而不是在抽象中生成最佳预测的技术背景中。如果要使用风险评估,它们必须基于对其现实影响的严格评估,而不是基于其理论潜力。
{"title":"Disparate Interactions: An Algorithm-in-the-Loop Analysis of Fairness in Risk Assessments","authors":"Ben Green, Yiling Chen","doi":"10.1145/3287560.3287563","DOIUrl":"https://doi.org/10.1145/3287560.3287563","url":null,"abstract":"Despite vigorous debates about the technical characteristics of risk assessments being deployed in the U.S. criminal justice system, remarkably little research has studied how these tools affect actual decision-making processes. After all, risk assessments do not make definitive decisions---they inform judges, who are the final arbiters. It is therefore essential that considerations of risk assessments be informed by rigorous studies of how judges actually interpret and use them. This paper takes a first step toward such research on human interactions with risk assessments through a controlled experimental study on Amazon Mechanical Turk. We found several behaviors that call into question the supposed efficacy and fairness of risk assessments: our study participants 1) underperformed the risk assessment even when presented with its predictions, 2) could not effectively evaluate the accuracy of their own or the risk assessment's predictions, and 3) exhibited behaviors fraught with \"disparate interactions,\" whereby the use of risk assessments led to higher risk predictions about black defendants and lower risk predictions about white defendants. These results suggest the need for a new \"algorithm-in-the-loop\" framework that places machine learning decision-making aids into the sociotechnical context of improving human decisions rather than the technical context of generating the best prediction in the abstract. If risk assessments are to be used at all, they must be grounded in rigorous evaluations of their real-world impacts instead of in their theoretical potential.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86888592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 200
Robot Eyes Wide Shut: Understanding Dishonest Anthropomorphism 机器人睁大眼睛:理解不诚实的拟人化
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287591
Brenda Leong, Evan Selinger
The goal of this paper is to advance design, policy, and ethics scholarship on how engineers and regulators can protect consumers from deceptive robots and artificial intelligences that exhibit the problem of dishonest anthropomorphism. The analysis expands upon ideas surrounding the principle of honest anthropomorphism originally formulated by Margot Kaminsky, Mathew Ruben, William D. Smart, and Cindy M. Grimm in their groundbreaking Maryland Law Review article, "Averting Robot Eyes." Applying boundary management theory and philosophical insights into prediction and perception, we create a new taxonomy that identifies fundamental types of dishonest anthropomorphism and pinpoints harms that they can cause. To demonstrate how the taxonomy can be applied as well as clarify the scope of the problems that it can cover, we critically consider a representative series of ethical issues, proposals, and questions concerning whether the principle of honest anthropomorphism has been violated.
本文的目标是推进设计、政策和伦理学术研究,探讨工程师和监管机构如何保护消费者免受欺骗性机器人和人工智能的侵害,这些机器人和人工智能表现出不诚实的拟人化问题。该分析扩展了围绕诚实拟人化原则的思想,该原则最初是由Margot Kaminsky, Mathew Ruben, William D. Smart和Cindy M. Grimm在他们开创性的马里兰法律评论文章“避开机器人的眼睛”中提出的。将边界管理理论和哲学见解应用到预测和感知中,我们创建了一种新的分类法,确定了不诚实的拟人化的基本类型,并指出了它们可能造成的危害。为了证明分类法可以如何应用,并澄清它可以涵盖的问题范围,我们批判性地考虑了一系列具有代表性的伦理问题、建议,以及关于诚实拟人化原则是否被违反的问题。
{"title":"Robot Eyes Wide Shut: Understanding Dishonest Anthropomorphism","authors":"Brenda Leong, Evan Selinger","doi":"10.1145/3287560.3287591","DOIUrl":"https://doi.org/10.1145/3287560.3287591","url":null,"abstract":"The goal of this paper is to advance design, policy, and ethics scholarship on how engineers and regulators can protect consumers from deceptive robots and artificial intelligences that exhibit the problem of dishonest anthropomorphism. The analysis expands upon ideas surrounding the principle of honest anthropomorphism originally formulated by Margot Kaminsky, Mathew Ruben, William D. Smart, and Cindy M. Grimm in their groundbreaking Maryland Law Review article, \"Averting Robot Eyes.\" Applying boundary management theory and philosophical insights into prediction and perception, we create a new taxonomy that identifies fundamental types of dishonest anthropomorphism and pinpoints harms that they can cause. To demonstrate how the taxonomy can be applied as well as clarify the scope of the problems that it can cover, we critically consider a representative series of ethical issues, proposals, and questions concerning whether the principle of honest anthropomorphism has been violated.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83127589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Fair Allocation through Competitive Equilibrium from Generic Incomes 从一般收入看竞争均衡下的公平分配
Pub Date : 2019-01-29 DOI: 10.1145/3287560.3287582
Moshe Babaioff, N. Nisan, Inbal Talgam-Cohen
Two food banks catering to populations of different sizes with different needs must divide among themselves a donation of food items. What constitutes a "fair" allocation of the items among them? Competitive equilibrium from equal incomes (CEEI) is a classic solution to the problem of fair and efficient allocation of goods among equally entitled agents [Foley 1967, Varian 1974]. Every agent (foodbank) receives an equal endowment of artificial currency with which to "purchase" bundles of goods (food items). Prices for the goods are set high enough such that the agents can simultaneously get their favorite within-budget bundle, and low enough such that all goods are allocated (no waste). A CEEI satisfies mathematical notions of fairness like fair-share, and also has built-in transparency -- prices can be published so the agents can verify they're being treated equally. However, a CEEI is not guaranteed to exist when the items are indivisible. We study competitive equilibrium from generic incomes (CEGI), which is based on the idea of slightly perturbed endowments, and enjoys similar fairness, efficiency and transparency properties as CEEI. We show that when the two agents have almost equal endowments and additive preferences for the items, a CEGI always exists. We then consider agents who are a priori non-equal (like different-sized foodbanks); we formulate a new notion of fair allocation among non-equals satisfied by CEGI, and show existence in cases of interest (like when the agents have identical preferences). Experiments on simulated and Spliddit data (a popular fair division website) indicate more general existence. Our results open opportunities for future research on fairness through generic endowments, and on fair treatment of non-equals.
两个食物银行为不同规模、不同需求的人口提供服务,它们之间必须分配捐赠的食物。怎样才能“公平”地分配这些物品呢?平等收入的竞争均衡(CEEI)是解决平等权利主体之间公平有效分配商品问题的经典方法[Foley 1967, Varian 1974]。每个代理(食物库)都获得相同的人工货币,用来“购买”成捆的商品(食物)。商品的价格设定得足够高,这样代理们就可以同时在预算范围内获得他们最喜欢的商品,同时又设定得足够低,这样所有的商品都能得到分配(没有浪费)。CEEI满足公平的数学概念,比如公平分享,而且还具有内置的透明度——价格可以公布,这样代理就可以验证他们受到了平等对待。但是,当项目不可分割时,不能保证存在CEEI。本文从一般收入(CEGI)的角度研究竞争均衡,它基于微扰动禀赋的思想,具有与一般收入相似的公平性、效率和透明度。我们证明了当两个智能体对物品具有几乎相等的禀赋和附加偏好时,一个CEGI总是存在的。然后我们考虑先天不相等的代理(比如不同规模的食物银行);我们提出了一个新的概念,即在CEGI满足的不平等之间公平分配,并证明了利益情况下的存在性(比如当代理具有相同的偏好时)。在模拟和Spliddit(一个流行的公平划分网站)数据上的实验表明,这种情况更为普遍。我们的研究结果为未来研究通用捐赠的公平性以及对不平等者的公平待遇提供了机会。
{"title":"Fair Allocation through Competitive Equilibrium from Generic Incomes","authors":"Moshe Babaioff, N. Nisan, Inbal Talgam-Cohen","doi":"10.1145/3287560.3287582","DOIUrl":"https://doi.org/10.1145/3287560.3287582","url":null,"abstract":"Two food banks catering to populations of different sizes with different needs must divide among themselves a donation of food items. What constitutes a \"fair\" allocation of the items among them? Competitive equilibrium from equal incomes (CEEI) is a classic solution to the problem of fair and efficient allocation of goods among equally entitled agents [Foley 1967, Varian 1974]. Every agent (foodbank) receives an equal endowment of artificial currency with which to \"purchase\" bundles of goods (food items). Prices for the goods are set high enough such that the agents can simultaneously get their favorite within-budget bundle, and low enough such that all goods are allocated (no waste). A CEEI satisfies mathematical notions of fairness like fair-share, and also has built-in transparency -- prices can be published so the agents can verify they're being treated equally. However, a CEEI is not guaranteed to exist when the items are indivisible. We study competitive equilibrium from generic incomes (CEGI), which is based on the idea of slightly perturbed endowments, and enjoys similar fairness, efficiency and transparency properties as CEEI. We show that when the two agents have almost equal endowments and additive preferences for the items, a CEGI always exists. We then consider agents who are a priori non-equal (like different-sized foodbanks); we formulate a new notion of fair allocation among non-equals satisfied by CEGI, and show existence in cases of interest (like when the agents have identical preferences). Experiments on simulated and Spliddit data (a popular fair division website) indicate more general existence. Our results open opportunities for future research on fairness through generic endowments, and on fair treatment of non-equals.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79528978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Bias in Bios: A Case Study of Semantic Representation Bias in a High-Stakes Setting Bios中的偏见:一个高风险情境下语义表示偏见的案例研究
Pub Date : 2019-01-27 DOI: 10.1145/3287560.3287572
Maria De-Arteaga, Alexey Romanov, Hanna M. Wallach, J. Chayes, C. Borgs, A. Chouldechova, S. Geyik, K. Kenthapadi, A. Kalai
We present a large-scale study of gender bias in occupation classification, a task where the use of machine learning may lead to negative outcomes on peoples' lives. We analyze the potential allocation harms that can result from semantic representation bias. To do so, we study the impact on occupation classification of including explicit gender indicators---such as first names and pronouns---in different semantic representations of online biographies. Additionally, we quantify the bias that remains when these indicators are "scrubbed," and describe proxy behavior that occurs in the absence of explicit gender indicators. As we demonstrate, differences in true positive rates between genders are correlated with existing gender imbalances in occupations, which may compound these imbalances.
我们提出了一项关于职业分类中性别偏见的大规模研究,在这项任务中,机器学习的使用可能会对人们的生活产生负面影响。我们分析了语义表示偏差可能导致的分配危害。为此,我们研究了在网络传记的不同语义表示中包含明确的性别指标(如名字和代词)对职业分类的影响。此外,我们量化了当这些指标被“抹掉”时仍然存在的偏见,并描述在没有明确性别指标的情况下发生的代理行为。正如我们所证明的那样,性别之间真实阳性率的差异与职业中现有的性别不平衡有关,这可能会加剧这些不平衡。
{"title":"Bias in Bios: A Case Study of Semantic Representation Bias in a High-Stakes Setting","authors":"Maria De-Arteaga, Alexey Romanov, Hanna M. Wallach, J. Chayes, C. Borgs, A. Chouldechova, S. Geyik, K. Kenthapadi, A. Kalai","doi":"10.1145/3287560.3287572","DOIUrl":"https://doi.org/10.1145/3287560.3287572","url":null,"abstract":"We present a large-scale study of gender bias in occupation classification, a task where the use of machine learning may lead to negative outcomes on peoples' lives. We analyze the potential allocation harms that can result from semantic representation bias. To do so, we study the impact on occupation classification of including explicit gender indicators---such as first names and pronouns---in different semantic representations of online biographies. Additionally, we quantify the bias that remains when these indicators are \"scrubbed,\" and describe proxy behavior that occurs in the absence of explicit gender indicators. As we demonstrate, differences in true positive rates between genders are correlated with existing gender imbalances in occupations, which may compound these imbalances.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74301859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 304
期刊
Proceedings of the Conference on Fairness, Accountability, and Transparency
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1