针对中毒攻击的隐私保护联合学习调查

Feng Xia, Wenhao Cheng
{"title":"针对中毒攻击的隐私保护联合学习调查","authors":"Feng Xia, Wenhao Cheng","doi":"10.1007/s10586-024-04629-7","DOIUrl":null,"url":null,"abstract":"<p>Federated learning (FL) is designed to protect privacy of participants by not allowing direct access to the participants’ local datasets and training processes. This limitation hinders the server’s ability to verify the authenticity of the model updates sent by participants, making FL vulnerable to poisoning attacks. In addition, gradients in FL process can reveal private information about the local dataset of the participants. However, there is a contradiction between improving robustness against poisoning attacks and preserving privacy of participants. Privacy-preserving techniques aim to make their data indistinguishable from each other, which hinders the detection of abnormal data based on similarity. It is challenging to enhance both aspects simultaneously. The growing concern for data security and privacy protection has inspired us to undertake this research and compile this survey. In this survey, we investigate existing privacy-preserving defense strategies against poisoning attacks in FL. First, we introduce two important classifications of poisoning attacks: data poisoning attack and model poisoning attack. Second, we study plaintext-based defense strategies and classify them into two categories: poisoning tolerance and poisoning detection. Third, we investigate how the combination of privacy techniques and traditional detection strategies can be achieved to defend against poisoning attacks while protecting the privacy of the participants. Finally, we also discuss the challenges faced in the area of security and privacy.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A survey on privacy-preserving federated learning against poisoning attacks\",\"authors\":\"Feng Xia, Wenhao Cheng\",\"doi\":\"10.1007/s10586-024-04629-7\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Federated learning (FL) is designed to protect privacy of participants by not allowing direct access to the participants’ local datasets and training processes. This limitation hinders the server’s ability to verify the authenticity of the model updates sent by participants, making FL vulnerable to poisoning attacks. In addition, gradients in FL process can reveal private information about the local dataset of the participants. However, there is a contradiction between improving robustness against poisoning attacks and preserving privacy of participants. Privacy-preserving techniques aim to make their data indistinguishable from each other, which hinders the detection of abnormal data based on similarity. It is challenging to enhance both aspects simultaneously. The growing concern for data security and privacy protection has inspired us to undertake this research and compile this survey. In this survey, we investigate existing privacy-preserving defense strategies against poisoning attacks in FL. First, we introduce two important classifications of poisoning attacks: data poisoning attack and model poisoning attack. Second, we study plaintext-based defense strategies and classify them into two categories: poisoning tolerance and poisoning detection. Third, we investigate how the combination of privacy techniques and traditional detection strategies can be achieved to defend against poisoning attacks while protecting the privacy of the participants. Finally, we also discuss the challenges faced in the area of security and privacy.</p>\",\"PeriodicalId\":501576,\"journal\":{\"name\":\"Cluster Computing\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Cluster Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/s10586-024-04629-7\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s10586-024-04629-7","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

联合学习(FL)旨在保护参与者的隐私,不允许直接访问参与者的本地数据集和训练过程。这一限制阻碍了服务器验证参与者发送的模型更新真实性的能力,从而使 FL 容易受到中毒攻击。此外,FL 过程中的梯度会泄露参与者本地数据集的私人信息。然而,提高对中毒攻击的鲁棒性与保护参与者隐私之间存在矛盾。保护隐私的技术旨在使数据彼此不可区分,这就阻碍了基于相似性的异常数据检测。要同时增强这两方面的能力是一项挑战。人们对数据安全和隐私保护的关注与日俱增,这促使我们开展这项研究,并编写了这份调查报告。在本调查中,我们研究了 FL 中现有的针对中毒攻击的隐私保护防御策略。首先,我们介绍了两种重要的中毒攻击分类:数据中毒攻击和模型中毒攻击。其次,我们研究了基于明文的防御策略,并将其分为两类:中毒容忍和中毒检测。第三,我们研究了如何将隐私技术和传统检测策略结合起来,在保护参与者隐私的同时防御中毒攻击。最后,我们还讨论了在安全和隐私领域面临的挑战。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

摘要图片

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
A survey on privacy-preserving federated learning against poisoning attacks

Federated learning (FL) is designed to protect privacy of participants by not allowing direct access to the participants’ local datasets and training processes. This limitation hinders the server’s ability to verify the authenticity of the model updates sent by participants, making FL vulnerable to poisoning attacks. In addition, gradients in FL process can reveal private information about the local dataset of the participants. However, there is a contradiction between improving robustness against poisoning attacks and preserving privacy of participants. Privacy-preserving techniques aim to make their data indistinguishable from each other, which hinders the detection of abnormal data based on similarity. It is challenging to enhance both aspects simultaneously. The growing concern for data security and privacy protection has inspired us to undertake this research and compile this survey. In this survey, we investigate existing privacy-preserving defense strategies against poisoning attacks in FL. First, we introduce two important classifications of poisoning attacks: data poisoning attack and model poisoning attack. Second, we study plaintext-based defense strategies and classify them into two categories: poisoning tolerance and poisoning detection. Third, we investigate how the combination of privacy techniques and traditional detection strategies can be achieved to defend against poisoning attacks while protecting the privacy of the participants. Finally, we also discuss the challenges faced in the area of security and privacy.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Quantitative and qualitative similarity measure for data clustering analysis OntoXAI: a semantic web rule language approach for explainable artificial intelligence Multi-threshold image segmentation using a boosted whale optimization: case study of breast invasive ductal carcinomas PSO-ACO-based bi-phase lightweight intrusion detection system combined with GA optimized ensemble classifiers A scalable and power efficient MAC protocol with adaptive TDMA for M2M communication
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1