Osama Hanna;Antonious M. Girgis;Christina Fragouli;Suhas Diggavi
{"title":"Differentially Private Stochastic Linear Bandits: (Almost) for Free","authors":"Osama Hanna;Antonious M. Girgis;Christina Fragouli;Suhas Diggavi","doi":"10.1109/JSAIT.2024.3389954","DOIUrl":null,"url":null,"abstract":"In this paper, we propose differentially private algorithms for the problem of stochastic linear bandits in the central, local and shuffled models. In the central model, we achieve almost the same regret as the optimal non-private algorithms, which means we get privacy for free. In particular, we achieve a regret of \n<inline-formula> <tex-math>$\\tilde {O}\\left({\\sqrt {T}+{}\\frac {1}{\\varepsilon }}\\right)$ </tex-math></inline-formula>\n matching the known lower bound for private linear bandits, while the best previously known algorithm achieves \n<inline-formula> <tex-math>$\\tilde {O}\\left({{}\\frac {1}{\\varepsilon }\\sqrt {T}}\\right)$ </tex-math></inline-formula>\n. In the local case, we achieve a regret of \n<inline-formula> <tex-math>$\\tilde {O}\\left({{}\\frac {1}{\\varepsilon }{\\sqrt {T}}}\\right)$ </tex-math></inline-formula>\n which matches the non-private regret for constant \n<inline-formula> <tex-math>$\\varepsilon $ </tex-math></inline-formula>\n, but suffers a regret penalty when \n<inline-formula> <tex-math>$\\varepsilon $ </tex-math></inline-formula>\n is small. In the shuffled model, we also achieve regret of \n<inline-formula> <tex-math>$\\tilde {O}\\left({\\sqrt {T}+{}\\frac {1}{\\varepsilon }}\\right)$ </tex-math></inline-formula>\n while the best previously known algorithm suffers a regret of \n<inline-formula> <tex-math>$\\tilde {O}\\left({{}\\frac {1}{\\varepsilon }{T^{3/5}}}\\right)$ </tex-math></inline-formula>\n. Our numerical evaluation validates our theoretical results. Our results generalize for contextual linear bandits with known context distributions.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"135-147"},"PeriodicalIF":0.0000,"publicationDate":"2024-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE journal on selected areas in information theory","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10502314/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we propose differentially private algorithms for the problem of stochastic linear bandits in the central, local and shuffled models. In the central model, we achieve almost the same regret as the optimal non-private algorithms, which means we get privacy for free. In particular, we achieve a regret of
$\tilde {O}\left({\sqrt {T}+{}\frac {1}{\varepsilon }}\right)$
matching the known lower bound for private linear bandits, while the best previously known algorithm achieves
$\tilde {O}\left({{}\frac {1}{\varepsilon }\sqrt {T}}\right)$
. In the local case, we achieve a regret of
$\tilde {O}\left({{}\frac {1}{\varepsilon }{\sqrt {T}}}\right)$
which matches the non-private regret for constant
$\varepsilon $
, but suffers a regret penalty when
$\varepsilon $
is small. In the shuffled model, we also achieve regret of
$\tilde {O}\left({\sqrt {T}+{}\frac {1}{\varepsilon }}\right)$
while the best previously known algorithm suffers a regret of
$\tilde {O}\left({{}\frac {1}{\varepsilon }{T^{3/5}}}\right)$
. Our numerical evaluation validates our theoretical results. Our results generalize for contextual linear bandits with known context distributions.