Dirichlet uncertainty wrappers for actionable algorithm accuracy accountability and auditability

José Mena Roldán, O. Vila, J. V. Marca
{"title":"Dirichlet uncertainty wrappers for actionable algorithm accuracy accountability and auditability","authors":"José Mena Roldán, O. Vila, J. V. Marca","doi":"10.1145/3351095.3372825","DOIUrl":null,"url":null,"abstract":"Nowadays, the use of machine learning models is becoming a utility in many applications. Companies deliver pre-trained models encapsulated as application programming interfaces (APIs) that developers combine with third-party components and their own models and data to create complex data products to solve specific problems. The complexity of such products and the lack of control and knowledge of the internals of each component used unavoidable cause effects, such as lack of transparency, difficulty in auditability, and the emergence of potential uncontrolled risks. They are effectively black-boxes. Accountability of such solutions is a challenge for the auditors and the machine learning community. In this work, we propose a wrapper that given a black-box model enriches its output prediction with a measure of uncertainty when applied to a target domain. To develop the wrapper, we follow these steps: Modeling the distribution of the output. In a text classification setting, the output is a probability distribution p(y|X, w*) over the different classes to predict, y, given an input text X and the pre-trained model with parameters w*. We model this output by a random variable to measure the variability that the data noise causes in the output. Here we consider the output distribution coming from a Dirichlet probability density function, thus p(y|X, w*) ~ Dir(α). Decomposition of the Dirichlet concentration parameter. To relate the output of the classifier with the concentration parameter in the Dirichlet distribution, we propose a decomposition of the concentration parameter in two terms: α = βy. The role of this scalar β is to control the spread of the distribution around the expected value, i.e. the original prediction y. Training the wrapper. Sentences are represented as the average value of their word embeddings. This representation feeds a neural network that outputs a single regression value that models the parameter β. For each input, we combine β and the black-box prediction to obtain the corresponding distribution for the output ym,i ~ Dir(αi). By using Monte Carlo sampling, we approximate the expected value of the classification probabilities, [EQUATION] and we train the model applying a cross-entropy loss over the predictions and the labels. Obtaining an uncertainty score from the wrapper. To obtain a numerical value for the uncertainty of a prediction, we draw samples from the resulting Dir(α) to evaluate the predictive entropy with [EQUATION], thus obtaining a numerical score for the uncertainty of each prediction. Using uncertainty for rejection. Based on this wrapper, we provide an actionable mechanism to mitigate risk in the form of decision rejection: once equipped with a value for the uncertainty of a given prediction, we can choose not to issue that prediction when the risk or uncertainty in that decision is significant. This results in a rejection system that selects the more confident predictions, discards those more uncertain, and leads to an improvement in the trustability of the resulting system. We showcase the proposed technique and methodology in a practical scenario where we apply a simulated sentiment analysis API based on NLP to different domains. On each experiment, we train a sentiment classifier using text reviews of products in a source domain. We apply the pre-trained black-box to obtain the predictions for the reviews from a target domain. The tuples of review plus black-box predictions are then used for training the wrapper to obtain the uncertainty. Finally, we use the uncertainty score to sort the predictions from more to less uncertain, and we search for a rejection point that maximizes the three performance measures: non-rejected accuracy, and classification and rejection quality. Experiments demonstrate the effectiveness of the uncertainty measure computed by the wrapper and shows its high correlation to bad quality predictions and misclassifications. In all the cases, the uncertainty metric here proposed outperforms traditional uncertainty measures.","PeriodicalId":377829,"journal":{"name":"Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency","volume":"142 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3351095.3372825","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

Nowadays, the use of machine learning models is becoming a utility in many applications. Companies deliver pre-trained models encapsulated as application programming interfaces (APIs) that developers combine with third-party components and their own models and data to create complex data products to solve specific problems. The complexity of such products and the lack of control and knowledge of the internals of each component used unavoidable cause effects, such as lack of transparency, difficulty in auditability, and the emergence of potential uncontrolled risks. They are effectively black-boxes. Accountability of such solutions is a challenge for the auditors and the machine learning community. In this work, we propose a wrapper that given a black-box model enriches its output prediction with a measure of uncertainty when applied to a target domain. To develop the wrapper, we follow these steps: Modeling the distribution of the output. In a text classification setting, the output is a probability distribution p(y|X, w*) over the different classes to predict, y, given an input text X and the pre-trained model with parameters w*. We model this output by a random variable to measure the variability that the data noise causes in the output. Here we consider the output distribution coming from a Dirichlet probability density function, thus p(y|X, w*) ~ Dir(α). Decomposition of the Dirichlet concentration parameter. To relate the output of the classifier with the concentration parameter in the Dirichlet distribution, we propose a decomposition of the concentration parameter in two terms: α = βy. The role of this scalar β is to control the spread of the distribution around the expected value, i.e. the original prediction y. Training the wrapper. Sentences are represented as the average value of their word embeddings. This representation feeds a neural network that outputs a single regression value that models the parameter β. For each input, we combine β and the black-box prediction to obtain the corresponding distribution for the output ym,i ~ Dir(αi). By using Monte Carlo sampling, we approximate the expected value of the classification probabilities, [EQUATION] and we train the model applying a cross-entropy loss over the predictions and the labels. Obtaining an uncertainty score from the wrapper. To obtain a numerical value for the uncertainty of a prediction, we draw samples from the resulting Dir(α) to evaluate the predictive entropy with [EQUATION], thus obtaining a numerical score for the uncertainty of each prediction. Using uncertainty for rejection. Based on this wrapper, we provide an actionable mechanism to mitigate risk in the form of decision rejection: once equipped with a value for the uncertainty of a given prediction, we can choose not to issue that prediction when the risk or uncertainty in that decision is significant. This results in a rejection system that selects the more confident predictions, discards those more uncertain, and leads to an improvement in the trustability of the resulting system. We showcase the proposed technique and methodology in a practical scenario where we apply a simulated sentiment analysis API based on NLP to different domains. On each experiment, we train a sentiment classifier using text reviews of products in a source domain. We apply the pre-trained black-box to obtain the predictions for the reviews from a target domain. The tuples of review plus black-box predictions are then used for training the wrapper to obtain the uncertainty. Finally, we use the uncertainty score to sort the predictions from more to less uncertain, and we search for a rejection point that maximizes the three performance measures: non-rejected accuracy, and classification and rejection quality. Experiments demonstrate the effectiveness of the uncertainty measure computed by the wrapper and shows its high correlation to bad quality predictions and misclassifications. In all the cases, the uncertainty metric here proposed outperforms traditional uncertainty measures.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Dirichlet不确定性包装可操作的算法,准确性,可问责性和可审计性
如今,机器学习模型的使用正在成为许多应用程序中的实用工具。公司将预先训练好的模型封装为应用程序编程接口(api),开发人员将其与第三方组件以及他们自己的模型和数据相结合,以创建复杂的数据产品来解决特定问题。此类产品的复杂性以及对所使用的每个组件的内部控制和知识的缺乏不可避免地会产生影响,例如缺乏透明度,难以审计,以及出现潜在的不受控制的风险。它们实际上是黑盒子。这些解决方案的问责制对审计人员和机器学习社区来说是一个挑战。在这项工作中,我们提出了一种包装器,该包装器给出了一个黑箱模型,当应用于目标域时,它通过不确定性的度量来丰富其输出预测。要开发包装器,我们需要遵循以下步骤:对输出的分布建模。在文本分类设置中,输出是要预测的不同类的概率分布p(y|X, w*), y,给定输入文本X和参数为w*的预训练模型。我们通过一个随机变量来模拟这个输出,以测量数据噪声在输出中引起的可变性。这里我们考虑来自Dirichlet概率密度函数的输出分布,即p(y|X, w*) ~ Dir(α)。狄利克雷浓度参数的分解。为了将分类器的输出与Dirichlet分布中的浓度参数联系起来,我们提出了浓度参数的两项分解:α = βy。这个标量β的作用是控制期望值周围分布的扩散,即原始预测y。训练包装器。句子被表示为它们的词嵌入的平均值。这种表示提供给一个神经网络,该网络输出一个单一的回归值,该回归值对参数β建模。对于每个输入,我们结合β和黑箱预测得到输出ym,i ~ Dir(αi)的相应分布。通过使用蒙特卡罗采样,我们近似分类概率的期望值,[等式],我们在预测和标签上应用交叉熵损失来训练模型。从包装器获得不确定度评分。为了获得预测不确定性的数值,我们从得到的Dir(α)中抽取样本,用[EQUATION]来评估预测熵,从而得到每个预测的不确定性的数值得分。用不确定性来拒绝。基于这个包装器,我们提供了一种可操作的机制,以决策拒绝的形式减轻风险:一旦为给定预测的不确定性配备了一个值,当该决策中的风险或不确定性很重要时,我们可以选择不发布该预测。这导致了一个拒绝系统,它选择了更有信心的预测,丢弃了那些更不确定的预测,并导致了结果系统的可信度的提高。我们在一个实际场景中展示了所提出的技术和方法,我们将基于NLP的模拟情感分析API应用于不同的领域。在每个实验中,我们使用源域中产品的文本评论来训练情感分类器。我们应用预训练的黑盒从目标域获得评论的预测。然后使用回顾和黑盒预测的元组来训练包装器以获得不确定性。最后,我们使用不确定性分数对预测进行从多到少的不确定性排序,并寻找一个拒绝点,使三个性能度量最大化:非拒绝准确性,分类和拒绝质量。实验证明了包装器计算的不确定度度量的有效性,并表明其与不良预测和错误分类的高度相关。在所有情况下,本文提出的不确定性度量都优于传统的不确定性度量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Dirichlet uncertainty wrappers for actionable algorithm accuracy accountability and auditability Algorithmic targeting of social policies: fairness, accuracy, and distributed governance Regulating transparency?: Facebook, Twitter and the German Network Enforcement Act CtrlZ.AI zine fair: critical perspectives Fairness, accountability, transparency in AI at scale: lessons from national programs
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1