Enhanced Privacy Bound for Shuffle Model with Personalized Privacy

Yixuan Liu, Yuhan Liu, Li Xiong, Yujie Gu, Hong Chen
{"title":"Enhanced Privacy Bound for Shuffle Model with Personalized Privacy","authors":"Yixuan Liu, Yuhan Liu, Li Xiong, Yujie Gu, Hong Chen","doi":"arxiv-2407.18157","DOIUrl":null,"url":null,"abstract":"The shuffle model of Differential Privacy (DP) is an enhanced privacy\nprotocol which introduces an intermediate trusted server between local users\nand a central data curator. It significantly amplifies the central DP guarantee\nby anonymizing and shuffling the local randomized data. Yet, deriving a tight\nprivacy bound is challenging due to its complicated randomization protocol.\nWhile most existing work are focused on unified local privacy settings, this\nwork focuses on deriving the central privacy bound for a more practical setting\nwhere personalized local privacy is required by each user. To bound the privacy\nafter shuffling, we first need to capture the probability of each user\ngenerating clones of the neighboring data points. Second, we need to quantify\nthe indistinguishability between two distributions of the number of clones on\nneighboring datasets. Existing works either inaccurately capture the\nprobability, or underestimate the indistinguishability between neighboring\ndatasets. Motivated by this, we develop a more precise analysis, which yields a\ngeneral and tighter bound for arbitrary DP mechanisms. Firstly, we derive the\nclone-generating probability by hypothesis testing %from a randomizer-specific\nperspective, which leads to a more accurate characterization of the\nprobability. Secondly, we analyze the indistinguishability in the context of\n$f$-DP, where the convexity of the distributions is leveraged to achieve a\ntighter privacy bound. Theoretical and numerical results demonstrate that our\nbound remarkably outperforms the existing results in the literature.","PeriodicalId":501123,"journal":{"name":"arXiv - CS - Databases","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Databases","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.18157","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The shuffle model of Differential Privacy (DP) is an enhanced privacy protocol which introduces an intermediate trusted server between local users and a central data curator. It significantly amplifies the central DP guarantee by anonymizing and shuffling the local randomized data. Yet, deriving a tight privacy bound is challenging due to its complicated randomization protocol. While most existing work are focused on unified local privacy settings, this work focuses on deriving the central privacy bound for a more practical setting where personalized local privacy is required by each user. To bound the privacy after shuffling, we first need to capture the probability of each user generating clones of the neighboring data points. Second, we need to quantify the indistinguishability between two distributions of the number of clones on neighboring datasets. Existing works either inaccurately capture the probability, or underestimate the indistinguishability between neighboring datasets. Motivated by this, we develop a more precise analysis, which yields a general and tighter bound for arbitrary DP mechanisms. Firstly, we derive the clone-generating probability by hypothesis testing %from a randomizer-specific perspective, which leads to a more accurate characterization of the probability. Secondly, we analyze the indistinguishability in the context of $f$-DP, where the convexity of the distributions is leveraged to achieve a tighter privacy bound. Theoretical and numerical results demonstrate that our bound remarkably outperforms the existing results in the literature.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
增强洗牌模型的个性化隐私约束
差分隐私(DP)的洗牌模型是一种增强型隐私协议,它在本地用户和中央数据管理员之间引入了一个中间可信服务器。它通过对本地随机数据进行匿名化和洗牌,大大增强了中央 DP 保证。然而,由于其随机化协议复杂,要推导出严密的隐私约束具有挑战性。现有的大部分工作都集中在统一的本地隐私设置上,而本工作则侧重于推导出更实用的中央隐私约束,即每个用户都需要个性化的本地隐私。为了约束洗牌后的隐私,我们首先需要捕捉每个用户生成相邻数据点克隆的概率。其次,我们需要量化相邻数据集上克隆数量的两种分布之间的不可分性。现有的研究要么没有准确捕捉到这种概率,要么低估了相邻数据集之间的不可分性。受此启发,我们开发了一种更精确的分析方法,可为任意 DP 机制提供更宽泛、更严格的约束。首先,我们通过假设检验%,从特定随机化器的角度推导出了克隆产生概率,从而更准确地描述了该概率。其次,我们分析了$f$-DP背景下的不可区分性,利用分布的凸性实现了更高的隐私约束。理论和数值结果表明,我们的边界明显优于文献中的现有结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Development of Data Evaluation Benchmark for Data Wrangling Recommendation System Messy Code Makes Managing ML Pipelines Difficult? Just Let LLMs Rewrite the Code! Fast and Adaptive Bulk Loading of Multidimensional Points Matrix Profile for Anomaly Detection on Multidimensional Time Series Extending predictive process monitoring for collaborative processes
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1