Threats, Attacks, and Defenses in Machine Unlearning: A Survey

Ziyao Liu;Huanyi Ye;Chen Chen;Yongsen Zheng;Kwok-Yan Lam
{"title":"Threats, Attacks, and Defenses in Machine Unlearning: A Survey","authors":"Ziyao Liu;Huanyi Ye;Chen Chen;Yongsen Zheng;Kwok-Yan Lam","doi":"10.1109/OJCS.2025.3543483","DOIUrl":null,"url":null,"abstract":"Machine Unlearning (MU) has recently gained considerable attention due to its potential to achieve Safe AI by removing the influence of specific data from trained Machine Learning (ML) models. This process, known as knowledge removal, addresses AI governance concerns of training data such as quality, sensitivity, copyright restrictions, obsolescence, and compliance with privacy regulations. Efforts have been made to design efficient unlearning approaches, with MU services being examined for integration with existing machine learning as a service (MLaaS), allowing users to submit requests to remove specific data from the training corpus. However, recent research highlights vulnerabilities in machine unlearning systems that can lead to significant security and privacy concerns. Moreover, extensive research indicates that unlearning methods and prevalent attacks fulfill diverse roles within MU systems. This underscores the intricate relationship and complex interplay among these mechanisms in maintaining system functionality and safety. This survey aims to fill the gap between the extensive number of studies on threats, attacks, and defenses in machine unlearning and the absence of a comprehensive review that categorizes their taxonomy, methods, and solutions, thus offering valuable insights for future research directions and practical implementations.","PeriodicalId":13205,"journal":{"name":"IEEE Open Journal of the Computer Society","volume":"6 ","pages":"413-425"},"PeriodicalIF":0.0000,"publicationDate":"2025-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10892039","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Open Journal of the Computer Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10892039/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Machine Unlearning (MU) has recently gained considerable attention due to its potential to achieve Safe AI by removing the influence of specific data from trained Machine Learning (ML) models. This process, known as knowledge removal, addresses AI governance concerns of training data such as quality, sensitivity, copyright restrictions, obsolescence, and compliance with privacy regulations. Efforts have been made to design efficient unlearning approaches, with MU services being examined for integration with existing machine learning as a service (MLaaS), allowing users to submit requests to remove specific data from the training corpus. However, recent research highlights vulnerabilities in machine unlearning systems that can lead to significant security and privacy concerns. Moreover, extensive research indicates that unlearning methods and prevalent attacks fulfill diverse roles within MU systems. This underscores the intricate relationship and complex interplay among these mechanisms in maintaining system functionality and safety. This survey aims to fill the gap between the extensive number of studies on threats, attacks, and defenses in machine unlearning and the absence of a comprehensive review that categorizes their taxonomy, methods, and solutions, thus offering valuable insights for future research directions and practical implementations.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
机器学习中的威胁、攻击和防御:综述
机器学习(MU)最近获得了相当大的关注,因为它有可能通过消除训练有素的机器学习(ML)模型中特定数据的影响来实现安全的人工智能。这个过程被称为知识移除,它解决了训练数据的人工智能治理问题,如质量、敏感性、版权限制、过时和遵守隐私法规。已经努力设计有效的取消学习方法,正在检查MU服务是否与现有的机器学习即服务(MLaaS)集成,允许用户提交请求以从训练语料库中删除特定数据。然而,最近的研究强调了机器学习系统中的漏洞,这些漏洞可能导致严重的安全和隐私问题。此外,广泛的研究表明,遗忘方法和流行的攻击在MU系统中扮演着不同的角色。这强调了维护系统功能和安全的这些机制之间错综复杂的关系和复杂的相互作用。本调查旨在填补大量关于机器学习中威胁、攻击和防御的研究与缺乏对其分类、方法和解决方案进行分类的全面审查之间的空白,从而为未来的研究方向和实际实施提供有价值的见解。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
12.60
自引率
0.00%
发文量
0
期刊最新文献
Integrating AI and Large Language Models for Automated Data Quality Enhancement in Data Integration Systems Reinforcing CBDC Integrity: A Novel Anti Money Laundering Solution by Integrating Blockchain, Machine Learning, and Taint Analysis Advancing Autism Spectrum Disorder Diagnosis With Multimodal Data: A Survey MCP-Secure: A Runtime Access Control Layer for Privilege-Aware LLM Agent Tooling Risk Model for Choosing Signature Thresholds for Software Updates
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1