{"title":"Privately Learning Smooth Distributions on the Hypercube by Projections","authors":"Clément LalanneTSE-R, Sébastien GadatTSE-R, IUF","doi":"arxiv-2409.10083","DOIUrl":null,"url":null,"abstract":"Fueled by the ever-increasing need for statistics that guarantee the privacy\nof their training sets, this article studies the centrally-private estimation\nof Sobolev-smooth densities of probability over the hypercube in dimension d.\nThe contributions of this article are two-fold : Firstly, it generalizes the\none dimensional results of (Lalanne et al., 2023) to non-integer levels of\nsmoothness and to a high-dimensional setting, which is important for two\nreasons : it is more suited for modern learning tasks, and it allows\nunderstanding the relations between privacy, dimensionality and smoothness,\nwhich is a central question with differential privacy. Secondly, this article\npresents a private strategy of estimation that is data-driven (usually referred\nto as adaptive in Statistics) in order to privately choose an estimator that\nachieves a good bias-variance trade-off among a finite family of private\nprojection estimators without prior knowledge of the ground-truth smoothness\n$\\beta$. This is achieved by adapting the Lepskii method for private selection,\nby adding a new penalization term that makes the estimation privacy-aware.","PeriodicalId":501379,"journal":{"name":"arXiv - STAT - Statistics Theory","volume":"39 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10083","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Fueled by the ever-increasing need for statistics that guarantee the privacy
of their training sets, this article studies the centrally-private estimation
of Sobolev-smooth densities of probability over the hypercube in dimension d.
The contributions of this article are two-fold : Firstly, it generalizes the
one dimensional results of (Lalanne et al., 2023) to non-integer levels of
smoothness and to a high-dimensional setting, which is important for two
reasons : it is more suited for modern learning tasks, and it allows
understanding the relations between privacy, dimensionality and smoothness,
which is a central question with differential privacy. Secondly, this article
presents a private strategy of estimation that is data-driven (usually referred
to as adaptive in Statistics) in order to privately choose an estimator that
achieves a good bias-variance trade-off among a finite family of private
projection estimators without prior knowledge of the ground-truth smoothness
$\beta$. This is achieved by adapting the Lepskii method for private selection,
by adding a new penalization term that makes the estimation privacy-aware.