On the Pitfalls of Label Differential Privacy

Andrés Muñoz
{"title":"On the Pitfalls of Label Differential Privacy","authors":"Andrés Muñoz","doi":"10.52591/202112077","DOIUrl":null,"url":null,"abstract":"We study the privacy limitations of label differential privacy, which has emerged as an intermediate trust model between local and central differential privacy, where only the label of each training example is protected (and the features are assumed to be public). We show that the guarantees provided by label DP are significantly weaker than they appear, as an adversary can \"un-noise\" the perturbed labels. Formally we show that the privacy loss has a close connection with Jeffreys’ divergence of the conditional distribution between positive and negative labels, which allows explicit formulation of the trade-off between utility and privacy in this setting. Our results suggest how to select public features that optimize this trade-off. But we still show that there is no free lunch—instances where label differential privacy guarantees are strong are exactly those where a good classifier does not exist. We complement the negative results with a non-parametric estimator for the true privacy loss, and apply our techniques on large-scale benchmark data to demonstrate how to achieve a desired privacy protection.","PeriodicalId":355096,"journal":{"name":"LatinX in AI at Neural Information Processing Systems Conference 2021","volume":"335 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"LatinX in AI at Neural Information Processing Systems Conference 2021","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.52591/202112077","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

We study the privacy limitations of label differential privacy, which has emerged as an intermediate trust model between local and central differential privacy, where only the label of each training example is protected (and the features are assumed to be public). We show that the guarantees provided by label DP are significantly weaker than they appear, as an adversary can "un-noise" the perturbed labels. Formally we show that the privacy loss has a close connection with Jeffreys’ divergence of the conditional distribution between positive and negative labels, which allows explicit formulation of the trade-off between utility and privacy in this setting. Our results suggest how to select public features that optimize this trade-off. But we still show that there is no free lunch—instances where label differential privacy guarantees are strong are exactly those where a good classifier does not exist. We complement the negative results with a non-parametric estimator for the true privacy loss, and apply our techniques on large-scale benchmark data to demonstrate how to achieve a desired privacy protection.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
论标签差分隐私的缺陷
我们研究了标签差分隐私的隐私限制,它已经成为局部和中心差分隐私之间的中间信任模型,其中只有每个训练样例的标签受到保护(并且假设特征是公开的)。我们表明,标签DP提供的保证比它们看起来的要弱得多,因为对手可以“去噪”受干扰的标签。形式上,我们表明隐私损失与Jeffreys关于正面和负面标签之间条件分布的分歧有密切的联系,这使得在这种情况下可以明确地表述效用和隐私之间的权衡。我们的结果建议如何选择优化这种权衡的公共特征。但是我们仍然证明了没有免费的午餐——标签差异隐私保证很强的实例恰恰是那些不存在好的分类器的实例。我们用真实隐私损失的非参数估计器来补充负面结果,并将我们的技术应用于大规模基准数据,以演示如何实现期望的隐私保护。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Vehicle Speed Estimation Using Computer Vision and Evolutionary Camera Calibration On the Pitfalls of Label Differential Privacy A Pharmacovigilance Application of Social Media Mining: An Ensemble Approach for Automated Classification and Extraction of Drug Flexible Learning of Sparse Neural Networks via Constrained L0 Regularizations Curating the Twitter Election Integrity Datasets for Better Online Troll Characterization
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1