首页 > 最新文献

Proceedings of the Conference on Fairness, Accountability, and Transparency最新文献

英文 中文
Problem Formulation and Fairness 问题提法与公平性
Pub Date : 2019-01-08 DOI: 10.1145/3287560.3287567
Samir Passi, Solon Barocas
Formulating data science problems is an uncertain and difficult process. It requires various forms of discretionary work to translate high-level objectives or strategic goals into tractable problems, necessitating, among other things, the identification of appropriate target variables and proxies. While these choices are rarely self-evident, normative assessments of data science projects often take them for granted, even though different translations can raise profoundly different ethical concerns. Whether we consider a data science project fair often has as much to do with the formulation of the problem as any property of the resulting model. Building on six months of ethnographic fieldwork with a corporate data science team---and channeling ideas from sociology and history of science, critical data studies, and early writing on knowledge discovery in databases---we describe the complex set of actors and activities involved in problem formulation. Our research demonstrates that the specification and operationalization of the problem are always negotiated and elastic, and rarely worked out with explicit normative considerations in mind. In so doing, we show that careful accounts of everyday data science work can help us better understand how and why data science problems are posed in certain ways---and why specific formulations prevail in practice, even in the face of what might seem like normatively preferable alternatives. We conclude by discussing the implications of our findings, arguing that effective normative interventions will require attending to the practical work of problem formulation.
制定数据科学问题是一个不确定和困难的过程。它需要各种形式的自由裁量工作,将高级别目标或战略目标转化为可处理的问题,除其他外,需要确定适当的目标变量和代理。虽然这些选择很少是不言而喻的,但对数据科学项目的规范性评估往往认为它们是理所当然的,尽管不同的翻译可能会引发截然不同的伦理问题。我们是否认为一个数据科学项目是公平的,往往与问题的表述有关,也与最终模型的任何属性有关。我们与一个企业数据科学团队一起进行了为期六个月的人种学实地考察,并从社会学和科学史、关键数据研究和数据库中知识发现的早期著述中汲取了一些想法,在此基础上,我们描述了问题制定过程中涉及的一系列复杂的参与者和活动。我们的研究表明,该问题的规范和操作化始终是协商和弹性的,很少考虑到明确的规范性因素。在这样做的过程中,我们表明,对日常数据科学工作的仔细描述可以帮助我们更好地理解数据科学问题是如何以及为什么以某些方式提出的——以及为什么特定的公式在实践中占上风,即使面对可能看起来像是规范上更可取的替代方案。最后,我们讨论了我们的研究结果的含义,认为有效的规范性干预将需要参与问题制定的实际工作。
{"title":"Problem Formulation and Fairness","authors":"Samir Passi, Solon Barocas","doi":"10.1145/3287560.3287567","DOIUrl":"https://doi.org/10.1145/3287560.3287567","url":null,"abstract":"Formulating data science problems is an uncertain and difficult process. It requires various forms of discretionary work to translate high-level objectives or strategic goals into tractable problems, necessitating, among other things, the identification of appropriate target variables and proxies. While these choices are rarely self-evident, normative assessments of data science projects often take them for granted, even though different translations can raise profoundly different ethical concerns. Whether we consider a data science project fair often has as much to do with the formulation of the problem as any property of the resulting model. Building on six months of ethnographic fieldwork with a corporate data science team---and channeling ideas from sociology and history of science, critical data studies, and early writing on knowledge discovery in databases---we describe the complex set of actors and activities involved in problem formulation. Our research demonstrates that the specification and operationalization of the problem are always negotiated and elastic, and rarely worked out with explicit normative considerations in mind. In so doing, we show that careful accounts of everyday data science work can help us better understand how and why data science problems are posed in certain ways---and why specific formulations prevail in practice, even in the face of what might seem like normatively preferable alternatives. We conclude by discussing the implications of our findings, arguing that effective normative interventions will require attending to the practical work of problem formulation.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75079985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 155
Efficient Search for Diverse Coherent Explanations 有效地寻找各种连贯的解释
Pub Date : 2019-01-02 DOI: 10.1145/3287560.3287569
Chris Russell
This paper proposes new search algorithms for counterfactual explanations based upon mixed integer programming. We are concerned with complex data in which variables may take any value from a contiguous range or an additional set of discrete states. We propose a novel set of constraints that we refer to as a "mixed polytope" and show how this can be used with an integer programming solver to efficiently find coherent counterfactual explanations i.e. solutions that are guaranteed to map back onto the underlying data structure, while avoiding the need for brute-force enumeration. We also look at the problem of diverse explanations and show how these can be generated within our framework.
提出了一种基于混合整数规划的反事实解释搜索算法。我们关心的是复杂数据,其中变量可以取连续范围内的任意值,也可以取一组额外的离散状态。我们提出了一组新的约束,我们称之为“混合多面体”,并展示了如何将其与整数规划求解器一起使用,以有效地找到连贯的反事实解释,即保证映射回底层数据结构的解决方案,同时避免了对暴力枚举的需要。我们还研究了各种解释的问题,并展示了如何在我们的框架内生成这些解释。
{"title":"Efficient Search for Diverse Coherent Explanations","authors":"Chris Russell","doi":"10.1145/3287560.3287569","DOIUrl":"https://doi.org/10.1145/3287560.3287569","url":null,"abstract":"This paper proposes new search algorithms for counterfactual explanations based upon mixed integer programming. We are concerned with complex data in which variables may take any value from a contiguous range or an additional set of discrete states. We propose a novel set of constraints that we refer to as a \"mixed polytope\" and show how this can be used with an integer programming solver to efficiently find coherent counterfactual explanations i.e. solutions that are guaranteed to map back onto the underlying data structure, while avoiding the need for brute-force enumeration. We also look at the problem of diverse explanations and show how these can be generated within our framework.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82588574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 187
Proceedings of the Conference on Fairness, Accountability, and Transparency 公平、问责和透明度会议论文集
{"title":"Proceedings of the Conference on Fairness, Accountability, and Transparency","authors":"","doi":"10.1145/3287560","DOIUrl":"https://doi.org/10.1145/3287560","url":null,"abstract":"","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72684473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
From Fair Decision Making To Social Equality 从公平决策到社会平等
Pub Date : 2018-12-07 DOI: 10.1145/3287560.3287599
Hussein Mozannar, Mesrob I. Ohannessian, N. Srebro
The study of fairness in intelligent decision systems has mostly ignored long-term influence on the underlying population. Yet fairness considerations (e.g. affirmative action) have often the implicit goal of achieving balance among groups within the population. The most basic notion of balance is eventual equality between the qualifications of the groups. How can we incorporate influence dynamics in decision making? How well do dynamics-oblivious fairness policies fare in terms of reaching equality? In this paper, we propose a simple yet revealing model that encompasses (1) a selection process where an institution chooses from multiple groups according to their qualifications so as to maximize an institutional utility and (2) dynamics that govern the evolution of the groups' qualifications according to the imposed policies. We focus on demographic parity as the formalism of affirmative action. We first give conditions under which an unconstrained policy reaches equality on its own. In this case, surprisingly, imposing demographic parity may break equality. When it doesn't, one would expect the additional constraint to reduce utility, however, we show that utility may in fact increase. In real world scenarios, unconstrained policies do not lead to equality. In such cases, we show that although imposing demographic parity may remedy it, there is a danger that groups settle at a worse set of qualifications. As a silver lining, we also identify when the constraint not only leads to equality, but also improves all groups. These cases and trade-offs are instrumental in determining when and how imposing demographic parity can be beneficial in selection processes, both for the institution and for society on the long run.
对智能决策系统公平性的研究大多忽略了对潜在人群的长期影响。然而,公平考虑(例如平权行动)往往隐含着实现人口中各群体之间平衡的目标。最基本的平衡概念是各组资格之间的最终平等。我们如何在决策中纳入影响动态?动态无关的公平政策在实现平等方面表现如何?在本文中,我们提出了一个简单但具有启发性的模型,该模型包含(1)一个选择过程,一个机构根据其资格从多个群体中进行选择,以最大化制度效用;(2)根据所施加的政策控制群体资格演变的动态。我们把人口平等作为平权行动的形式主义来关注。我们首先给出不受约束的政策自行达到平等的条件。在这种情况下,令人惊讶的是,强制实行人口平等可能会打破平等。当它不这样做时,人们会期望额外的约束会降低效用,然而,我们表明效用实际上可能会增加。在现实世界中,不受约束的政策不会导致平等。在这种情况下,我们表明,尽管实行人口平等可能会补救它,但存在一种危险,即群体满足于一套更糟糕的资格。作为一线希望,我们也会发现约束不仅会带来平等,还会改善所有群体。从长远来看,这些案例和权衡有助于确定何时以及如何在选择过程中对机构和社会有利。
{"title":"From Fair Decision Making To Social Equality","authors":"Hussein Mozannar, Mesrob I. Ohannessian, N. Srebro","doi":"10.1145/3287560.3287599","DOIUrl":"https://doi.org/10.1145/3287560.3287599","url":null,"abstract":"The study of fairness in intelligent decision systems has mostly ignored long-term influence on the underlying population. Yet fairness considerations (e.g. affirmative action) have often the implicit goal of achieving balance among groups within the population. The most basic notion of balance is eventual equality between the qualifications of the groups. How can we incorporate influence dynamics in decision making? How well do dynamics-oblivious fairness policies fare in terms of reaching equality? In this paper, we propose a simple yet revealing model that encompasses (1) a selection process where an institution chooses from multiple groups according to their qualifications so as to maximize an institutional utility and (2) dynamics that govern the evolution of the groups' qualifications according to the imposed policies. We focus on demographic parity as the formalism of affirmative action. We first give conditions under which an unconstrained policy reaches equality on its own. In this case, surprisingly, imposing demographic parity may break equality. When it doesn't, one would expect the additional constraint to reduce utility, however, we show that utility may in fact increase. In real world scenarios, unconstrained policies do not lead to equality. In such cases, we show that although imposing demographic parity may remedy it, there is a danger that groups settle at a worse set of qualifications. As a silver lining, we also identify when the constraint not only leads to equality, but also improves all groups. These cases and trade-offs are instrumental in determining when and how imposing demographic parity can be beneficial in selection processes, both for the institution and for society on the long run.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87176836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 86
Racial categories in machine learning 机器学习中的种族分类
Pub Date : 2018-11-28 DOI: 10.1145/3287560.3287575
Sebastian Benthall, Bruce D. Haynes
Controversies around race and machine learning have sparked debate among computer scientists over how to design machine learning systems that guarantee fairness. These debates rarely engage with how racial identity is embedded in our social experience, making for sociological and psychological complexity. This complexity challenges the paradigm of considering fairness to be a formal property of supervised learning with respect to protected personal attributes. Racial identity is not simply a personal subjective quality. For people labeled "Black" it is an ascribed political category that has consequences for social differentiation embedded in systemic patterns of social inequality achieved through both social and spatial segregation. In the United States, racial classification can best be understood as a system of inherently unequal status categories that places whites as the most privileged category while signifying the Negro/black category as stigmatized. Social stigma is reinforced through the unequal distribution of societal rewards and goods along racial lines that is reinforced by state, corporate, and civic institutions and practices. This creates a dilemma for society and designers: be blind to racial group disparities and thereby reify racialized social inequality by no longer measuring systemic inequality, or be conscious of racial categories in a way that itself reifies race. We propose a third option. By preceding group fairness interventions with unsupervised learning to dynamically detect patterns of segregation, machine learning systems can mitigate the root cause of social disparities, social segregation and stratification, without further anchoring status categories of disadvantage.
关于种族和机器学习的争议引发了计算机科学家关于如何设计保证公平的机器学习系统的争论。这些辩论很少涉及种族身份如何嵌入我们的社会经验,从而导致社会学和心理学的复杂性。这种复杂性挑战了将公平性视为监督学习的正式属性的范式,而不是受保护的个人属性。种族认同不仅仅是个人的主观品质。对于那些被贴上“黑人”标签的人来说,这是一个被赋予的政治类别,它对社会分化产生了影响,这种社会分化植根于通过社会和空间隔离实现的社会不平等的系统性模式中。在美国,种族分类最好被理解为一种本质上不平等的地位类别制度,将白人视为最特权的类别,而将黑人/黑人类别视为耻辱。国家、企业和民间机构和实践进一步强化了社会奖励和商品沿种族界线的不平等分配,从而加剧了社会耻辱。这给社会和设计师带来了两难境地:要么无视种族群体差异,不再衡量系统性不平等,从而使种族化的社会不平等具体化,要么意识到种族类别本身就是种族的具体化。我们提出第三种选择。通过预先使用无监督学习进行群体公平干预,以动态检测隔离模式,机器学习系统可以减轻社会差异、社会隔离和分层的根源,而无需进一步锚定劣势地位类别。
{"title":"Racial categories in machine learning","authors":"Sebastian Benthall, Bruce D. Haynes","doi":"10.1145/3287560.3287575","DOIUrl":"https://doi.org/10.1145/3287560.3287575","url":null,"abstract":"Controversies around race and machine learning have sparked debate among computer scientists over how to design machine learning systems that guarantee fairness. These debates rarely engage with how racial identity is embedded in our social experience, making for sociological and psychological complexity. This complexity challenges the paradigm of considering fairness to be a formal property of supervised learning with respect to protected personal attributes. Racial identity is not simply a personal subjective quality. For people labeled \"Black\" it is an ascribed political category that has consequences for social differentiation embedded in systemic patterns of social inequality achieved through both social and spatial segregation. In the United States, racial classification can best be understood as a system of inherently unequal status categories that places whites as the most privileged category while signifying the Negro/black category as stigmatized. Social stigma is reinforced through the unequal distribution of societal rewards and goods along racial lines that is reinforced by state, corporate, and civic institutions and practices. This creates a dilemma for society and designers: be blind to racial group disparities and thereby reify racialized social inequality by no longer measuring systemic inequality, or be conscious of racial categories in a way that itself reifies race. We propose a third option. By preceding group fairness interventions with unsupervised learning to dynamically detect patterns of segregation, machine learning systems can mitigate the root cause of social disparities, social segregation and stratification, without further anchoring status categories of disadvantage.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89860551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 96
Fairness Under Unawareness: Assessing Disparity When Protected Class Is Unobserved 无意识下的公平:评估受保护阶层未被注意时的差异
Pub Date : 2018-11-27 DOI: 10.1145/3287560.3287594
Jiahao Chen, Nathan Kallus, Xiaojie Mao, G. Svacha, Madeleine Udell
Assessing the fairness of a decision making system with respect to a protected class, such as gender or race, is challenging when class membership labels are unavailable. Probabilistic models for predicting the protected class based on observable proxies, such as surname and geolocation for race, are sometimes used to impute these missing labels for compliance assessments. Empirically, these methods are observed to exaggerate disparities, but the reason why is unknown. In this paper, we decompose the biases in estimating outcome disparity via threshold-based imputation into multiple interpretable bias sources, allowing us to explain when over- or underestimation occurs. We also propose an alternative weighted estimator that uses soft classification, and show that its bias arises simply from the conditional covariance of the outcome with the true class membership. Finally, we illustrate our results with numerical simulations and a public dataset of mortgage applications, using geolocation as a proxy for race. We confirm that the bias of threshold-based imputation is generally upward, but its magnitude varies strongly with the threshold chosen. Our new weighted estimator tends to have a negative bias that is much simpler to analyze and reason about.
在无法获得类成员标签的情况下,评估与受保护的类(如性别或种族)相关的决策制定系统的公平性是具有挑战性的。根据可观察的代理(如姓氏和种族的地理位置)预测受保护类别的概率模型,有时用于为符合性评估输入这些缺失的标签。从经验上看,这些方法被观察到会夸大差异,但其原因尚不清楚。在本文中,我们通过基于阈值的插值将估计结果差异的偏差分解为多个可解释的偏差源,使我们能够解释何时发生高估或低估。我们还提出了一种使用软分类的替代加权估计器,并表明其偏差仅仅来自结果与真实类隶属度的条件协方差。最后,我们用数值模拟和抵押贷款申请的公共数据集来说明我们的结果,使用地理位置作为种族的代理。我们证实,基于阈值的归算的偏差通常是向上的,但其大小随阈值的选择而变化很大。我们的新加权估计器倾向于有一个更容易分析和推理的负偏差。
{"title":"Fairness Under Unawareness: Assessing Disparity When Protected Class Is Unobserved","authors":"Jiahao Chen, Nathan Kallus, Xiaojie Mao, G. Svacha, Madeleine Udell","doi":"10.1145/3287560.3287594","DOIUrl":"https://doi.org/10.1145/3287560.3287594","url":null,"abstract":"Assessing the fairness of a decision making system with respect to a protected class, such as gender or race, is challenging when class membership labels are unavailable. Probabilistic models for predicting the protected class based on observable proxies, such as surname and geolocation for race, are sometimes used to impute these missing labels for compliance assessments. Empirically, these methods are observed to exaggerate disparities, but the reason why is unknown. In this paper, we decompose the biases in estimating outcome disparity via threshold-based imputation into multiple interpretable bias sources, allowing us to explain when over- or underestimation occurs. We also propose an alternative weighted estimator that uses soft classification, and show that its bias arises simply from the conditional covariance of the outcome with the true class membership. Finally, we illustrate our results with numerical simulations and a public dataset of mortgage applications, using geolocation as a proxy for race. We confirm that the bias of threshold-based imputation is generally upward, but its magnitude varies strongly with the threshold chosen. Our new weighted estimator tends to have a negative bias that is much simpler to analyze and reason about.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73797294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 176
50 Years of Test (Un)fairness: Lessons for Machine Learning 50年的测试公平性:机器学习的经验教训
Pub Date : 2018-11-25 DOI: 10.1145/3287560.3287600
B. Hutchinson, Margaret Mitchell
Quantitative definitions of what is unfair and what is fair have been introduced in multiple disciplines for well over 50 years, including in education, hiring, and machine learning. We trace how the notion of fairness has been defined within the testing communities of education and hiring over the past half century, exploring the cultural and social context in which different fairness definitions have emerged. In some cases, earlier definitions of fairness are similar or identical to definitions of fairness in current machine learning research, and foreshadow current formal work. In other cases, insights into what fairness means and how to measure it have largely gone overlooked. We compare past and current notions of fairness along several dimensions, including the fairness criteria, the focus of the criteria (e.g., a test, a model, or its use), the relationship of fairness to individuals, groups, and subgroups, and the mathematical method for measuring fairness (e.g., classification, regression). This work points the way towards future research and measurement of (un)fairness that builds from our modern understanding of fairness while incorporating insights from the past.
50多年来,不公平和公平的定量定义已经在多个学科中引入,包括教育、招聘和机器学习。我们追溯了在过去的半个世纪里,公平的概念是如何在教育和招聘的测试社区中被定义的,探索了不同公平定义出现的文化和社会背景。在某些情况下,早期的公平定义与当前机器学习研究中的公平定义相似或相同,并预示着当前的正式工作。在其他情况下,对公平意味着什么以及如何衡量公平的见解在很大程度上被忽视了。我们从几个方面比较了过去和现在的公平概念,包括公平标准、标准的焦点(例如,测试、模型或其使用)、公平与个人、群体和子群体的关系,以及衡量公平的数学方法(例如,分类、回归)。这项工作为未来的研究和衡量(非)公平指明了方向,这些研究和衡量建立在我们对公平的现代理解的基础上,同时结合了过去的见解。
{"title":"50 Years of Test (Un)fairness: Lessons for Machine Learning","authors":"B. Hutchinson, Margaret Mitchell","doi":"10.1145/3287560.3287600","DOIUrl":"https://doi.org/10.1145/3287560.3287600","url":null,"abstract":"Quantitative definitions of what is unfair and what is fair have been introduced in multiple disciplines for well over 50 years, including in education, hiring, and machine learning. We trace how the notion of fairness has been defined within the testing communities of education and hiring over the past half century, exploring the cultural and social context in which different fairness definitions have emerged. In some cases, earlier definitions of fairness are similar or identical to definitions of fairness in current machine learning research, and foreshadow current formal work. In other cases, insights into what fairness means and how to measure it have largely gone overlooked. We compare past and current notions of fairness along several dimensions, including the fairness criteria, the focus of the criteria (e.g., a test, a model, or its use), the relationship of fairness to individuals, groups, and subgroups, and the mathematical method for measuring fairness (e.g., classification, regression). This work points the way towards future research and measurement of (un)fairness that builds from our modern understanding of fairness while incorporating insights from the past.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88391349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 278
Equality of Voice: Towards Fair Representation in Crowdsourced Top-K Recommendations 话语权平等:实现众包Top-K推荐中的公平代表权
Pub Date : 2018-11-21 DOI: 10.1145/3287560.3287570
Abhijnan Chakraborty, Gourab K. Patro, Niloy Ganguly, K. Gummadi, P. Loiseau
To help their users to discover important items at a particular time, major websites like Twitter, Yelp, TripAdvisor or NYTimes provide Top-K recommendations (e.g., 10 Trending Topics, Top 5 Hotels in Paris or 10 Most Viewed News Stories), which rely on crowdsourced popularity signals to select the items. However, different sections of a crowd may have different preferences, and there is a large silent majority who do not explicitly express their opinion. Also, the crowd often consists of actors like bots, spammers, or people running orchestrated campaigns. Recommendation algorithms today largely do not consider such nuances, hence are vulnerable to strategic manipulation by small but hyper-active user groups. To fairly aggregate the preferences of all users while recommending top-K items, we borrow ideas from prior research on social choice theory, and identify a voting mechanism called Single Transferable Vote (STV) as having many of the fairness properties we desire in top-K item (s)elections. We develop an innovative mechanism to attribute preferences of silent majority which also make STV completely operational. We show the generalizability of our approach by implementing it on two different real-world datasets. Through extensive experimentation and comparison with state-of-the-art techniques, we show that our proposed approach provides maximum user satisfaction, and cuts down drastically on items disliked by most but hyper-actively promoted by a few users.
为了帮助用户在特定时间发现重要的项目,像Twitter, Yelp, TripAdvisor或NYTimes这样的主要网站提供Top- k推荐(例如,10个热门话题,巴黎前5家酒店或10个最受欢迎的新闻故事),它们依靠众包的流行信号来选择项目。然而,人群的不同部分可能有不同的偏好,并且有一个很大的沉默的大多数,他们不明确地表达自己的意见。此外,这群人通常由机器人、垃圾邮件发送者或精心策划的活动的人组成。今天的推荐算法在很大程度上没有考虑到这些细微差别,因此很容易受到少数但高度活跃的用户群体的战略操纵。在推荐top-K项目时,为了公平地汇总所有用户的偏好,我们借鉴了先前的社会选择理论研究,并确定了一种称为单一可转移投票(STV)的投票机制,该机制具有我们在top-K项目选举中期望的许多公平性属性。我们开发了一种创新的机制来属性沉默的大多数的偏好,这也使STV完全可操作。我们通过在两个不同的真实世界数据集上实现它来展示我们方法的泛化性。通过广泛的实验和与最先进技术的比较,我们表明我们提出的方法提供了最大的用户满意度,并大大减少了大多数用户不喜欢但少数用户过度积极推广的项目。
{"title":"Equality of Voice: Towards Fair Representation in Crowdsourced Top-K Recommendations","authors":"Abhijnan Chakraborty, Gourab K. Patro, Niloy Ganguly, K. Gummadi, P. Loiseau","doi":"10.1145/3287560.3287570","DOIUrl":"https://doi.org/10.1145/3287560.3287570","url":null,"abstract":"To help their users to discover important items at a particular time, major websites like Twitter, Yelp, TripAdvisor or NYTimes provide Top-K recommendations (e.g., 10 Trending Topics, Top 5 Hotels in Paris or 10 Most Viewed News Stories), which rely on crowdsourced popularity signals to select the items. However, different sections of a crowd may have different preferences, and there is a large silent majority who do not explicitly express their opinion. Also, the crowd often consists of actors like bots, spammers, or people running orchestrated campaigns. Recommendation algorithms today largely do not consider such nuances, hence are vulnerable to strategic manipulation by small but hyper-active user groups. To fairly aggregate the preferences of all users while recommending top-K items, we borrow ideas from prior research on social choice theory, and identify a voting mechanism called Single Transferable Vote (STV) as having many of the fairness properties we desire in top-K item (s)elections. We develop an innovative mechanism to attribute preferences of silent majority which also make STV completely operational. We show the generalizability of our approach by implementing it on two different real-world datasets. Through extensive experimentation and comparison with state-of-the-art techniques, we show that our proposed approach provides maximum user satisfaction, and cuts down drastically on items disliked by most but hyper-actively promoted by a few users.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85705867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
On Human Predictions with Explanations and Predictions of Machine Learning Models: A Case Study on Deception Detection 机器学习模型的解释和预测:欺骗检测的案例研究
Pub Date : 2018-11-19 DOI: 10.1145/3287560.3287590
Vivian Lai, Chenhao Tan
Humans are the final decision makers in critical tasks that involve ethical and legal concerns, ranging from recidivism prediction, to medical diagnosis, to fighting against fake news. Although machine learning models can sometimes achieve impressive performance in these tasks, these tasks are not amenable to full automation. To realize the potential of machine learning for improving human decisions, it is important to understand how assistance from machine learning models affects human performance and human agency. In this paper, we use deception detection as a testbed and investigate how we can harness explanations and predictions of machine learning models to improve human performance while retaining human agency. We propose a spectrum between full human agency and full automation, and develop varying levels of machine assistance along the spectrum that gradually increase the influence of machine predictions. We find that without showing predicted labels, explanations alone slightly improve human performance in the end task. In comparison, human performance is greatly improved by showing predicted labels (>20% relative improvement) and can be further improved by explicitly suggesting strong machine performance. Interestingly, when predicted labels are shown, explanations of machine predictions induce a similar level of accuracy as an explicit statement of strong machine performance. Our results demonstrate a tradeoff between human performance and human agency and show that explanations of machine predictions can moderate this tradeoff.
人类是涉及道德和法律问题的关键任务的最终决策者,从累犯预测到医疗诊断,再到打击假新闻。虽然机器学习模型有时可以在这些任务中取得令人印象深刻的表现,但这些任务不适合完全自动化。为了实现机器学习在改善人类决策方面的潜力,了解机器学习模型的帮助如何影响人类的表现和人类的代理是很重要的。在本文中,我们使用欺骗检测作为测试平台,并研究如何利用机器学习模型的解释和预测来提高人类的表现,同时保留人类的能动性。我们提出了完全人类代理和完全自动化之间的频谱,并沿着频谱开发不同水平的机器辅助,逐渐增加机器预测的影响。我们发现,在不显示预测标签的情况下,单独的解释会略微提高人类在最终任务中的表现。相比之下,通过显示预测的标签,人类的表现得到了极大的提高(>20%的相对提高),并且可以通过明确地提出强大的机器性能来进一步提高。有趣的是,当显示预测标签时,对机器预测的解释与对机器性能的明确陈述的准确性相似。我们的研究结果证明了人类表现和人类代理之间的权衡,并表明对机器预测的解释可以缓和这种权衡。
{"title":"On Human Predictions with Explanations and Predictions of Machine Learning Models: A Case Study on Deception Detection","authors":"Vivian Lai, Chenhao Tan","doi":"10.1145/3287560.3287590","DOIUrl":"https://doi.org/10.1145/3287560.3287590","url":null,"abstract":"Humans are the final decision makers in critical tasks that involve ethical and legal concerns, ranging from recidivism prediction, to medical diagnosis, to fighting against fake news. Although machine learning models can sometimes achieve impressive performance in these tasks, these tasks are not amenable to full automation. To realize the potential of machine learning for improving human decisions, it is important to understand how assistance from machine learning models affects human performance and human agency. In this paper, we use deception detection as a testbed and investigate how we can harness explanations and predictions of machine learning models to improve human performance while retaining human agency. We propose a spectrum between full human agency and full automation, and develop varying levels of machine assistance along the spectrum that gradually increase the influence of machine predictions. We find that without showing predicted labels, explanations alone slightly improve human performance in the end task. In comparison, human performance is greatly improved by showing predicted labels (>20% relative improvement) and can be further improved by explicitly suggesting strong machine performance. Interestingly, when predicted labels are shown, explanations of machine predictions induce a similar level of accuracy as an explicit statement of strong machine performance. Our results demonstrate a tradeoff between human performance and human agency and show that explanations of machine predictions can moderate this tradeoff.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85823061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 247
Deep Weighted Averaging Classifiers 深度加权平均分类器
Pub Date : 2018-11-06 DOI: 10.1145/3287560.3287595
Dallas Card, Michael J.Q. Zhang, Noah A. Smith
Recent advances in deep learning have achieved impressive gains in classification accuracy on a variety of types of data, including images and text. Despite these gains, however, concerns have been raised about the calibration, robustness, and interpretability of these models. In this paper we propose a simple way to modify any conventional deep architecture to automatically provide more transparent explanations for classification decisions, as well as an intuitive notion of the credibility of each prediction. Specifically, we draw on ideas from nonparametric kernel regression, and propose to predict labels based on a weighted sum of training instances, where the weights are determined by distance in a learned instance-embedding space. Working within the framework of conformal methods, we propose a new measure of nonconformity suggested by our model, and experimentally validate the accompanying theoretical expectations, demonstrating improved transparency, controlled error rates, and robustness to out-of-domain data, without compromising on accuracy or calibration.
深度学习的最新进展在各种类型的数据(包括图像和文本)的分类准确性方面取得了令人印象深刻的进展。然而,尽管取得了这些成果,但人们对这些模型的校准、稳健性和可解释性提出了担忧。在本文中,我们提出了一种简单的方法来修改任何传统的深度架构,以自动为分类决策提供更透明的解释,以及每个预测可信度的直观概念。具体来说,我们借鉴了非参数核回归的思想,并提出基于训练实例的加权和来预测标签,其中权重由学习到的实例嵌入空间中的距离决定。在共形方法的框架内,我们提出了一种由我们的模型提出的新的不一致性测量方法,并通过实验验证了伴随的理论期望,证明了改进的透明度,控制错误率和对域外数据的鲁棒性,而不影响准确性或校准。
{"title":"Deep Weighted Averaging Classifiers","authors":"Dallas Card, Michael J.Q. Zhang, Noah A. Smith","doi":"10.1145/3287560.3287595","DOIUrl":"https://doi.org/10.1145/3287560.3287595","url":null,"abstract":"Recent advances in deep learning have achieved impressive gains in classification accuracy on a variety of types of data, including images and text. Despite these gains, however, concerns have been raised about the calibration, robustness, and interpretability of these models. In this paper we propose a simple way to modify any conventional deep architecture to automatically provide more transparent explanations for classification decisions, as well as an intuitive notion of the credibility of each prediction. Specifically, we draw on ideas from nonparametric kernel regression, and propose to predict labels based on a weighted sum of training instances, where the weights are determined by distance in a learned instance-embedding space. Working within the framework of conformal methods, we propose a new measure of nonconformity suggested by our model, and experimentally validate the accompanying theoretical expectations, demonstrating improved transparency, controlled error rates, and robustness to out-of-domain data, without compromising on accuracy or calibration.","PeriodicalId":20573,"journal":{"name":"Proceedings of the Conference on Fairness, Accountability, and Transparency","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2018-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82451712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
期刊
Proceedings of the Conference on Fairness, Accountability, and Transparency
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1