{"title":"DPSGD Strategies for Cross-Silo Federated Learning","authors":"Matthieu Moreau, Tarek Benkhelif","doi":"10.1109/CCCI52664.2021.9583220","DOIUrl":null,"url":null,"abstract":"As federated learning (FL) grows and new techniques are created to improve its efficiency and robustness, differential privacy (DP) proved to be a good ally for protecting users’ information. The differentially private version of stochastic gradient descent (DPSGD) is one of the most promising methods for enforcing privacy in machine learning algorithms. The noise added in DPSGD plays an important role in the convergence and performance of a model but also in the resulting privacy guarantee and must thus be chosen carefully. This paper reviews the effects of either selecting fixed or adaptive noise when training federated models under the cross-silo setting. We highlight their strengths and weaknesses and propose a hybrid approach, getting the best of both worlds.","PeriodicalId":136382,"journal":{"name":"2021 International Conference on Communications, Computing, Cybersecurity, and Informatics (CCCI)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Communications, Computing, Cybersecurity, and Informatics (CCCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCCI52664.2021.9583220","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
As federated learning (FL) grows and new techniques are created to improve its efficiency and robustness, differential privacy (DP) proved to be a good ally for protecting users’ information. The differentially private version of stochastic gradient descent (DPSGD) is one of the most promising methods for enforcing privacy in machine learning algorithms. The noise added in DPSGD plays an important role in the convergence and performance of a model but also in the resulting privacy guarantee and must thus be chosen carefully. This paper reviews the effects of either selecting fixed or adaptive noise when training federated models under the cross-silo setting. We highlight their strengths and weaknesses and propose a hybrid approach, getting the best of both worlds.