{"title":"meMIA: Multilevel Ensemble Membership Inference Attack","authors":"Najeeb Ullah;Muhammad Naveed Aman;Biplab Sikdar","doi":"10.1109/TAI.2024.3445326","DOIUrl":null,"url":null,"abstract":"Leakage of private information in machine learning models can lead to breaches of confidentiality, identity theft, and unauthorized access to personal data. Ensuring the safe and trustworthy deployment of AI systems necessitates addressing privacy concerns to prevent unintentional disclosure and discrimination. One significant threat, membership inference (MI) attacks, exploit vulnerabilities in target learning models to determine if a given sample was part of the training set. However, the effectiveness of existing MI attacks is often limited by the number of classes in the dataset or the need for diverse multilevel adversarial features to exploit overfitted models. To enhance MI attack performance, we propose meMIA, a novel framework based on stacked ensemble learning. meMIA integrates embeddings from a neural network (NN) and a long short-term memory (LSTM) model, training a subsequent NN, termed the meta-model, on the concatenated embeddings. This method leverages the complementary strengths of NN and LSTM models; the LSTM captures order differences in confidence scores, while the NN discerns probability distribution differences between member and nonmember samples. We extensively evaluate meMIA on seven benchmark datasets, demonstrating that it surpasses current state-of-the-art MI attacks, achieving accuracy up to 94.6% and near-perfect recall. meMIA's superior performance, especially on datasets with fewer classes, underscores the urgent need for robust defenses against privacy attacks in machine learning, contributing to the safer and more ethical use of AI technologies.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 1","pages":"93-106"},"PeriodicalIF":0.0000,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10639374/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Leakage of private information in machine learning models can lead to breaches of confidentiality, identity theft, and unauthorized access to personal data. Ensuring the safe and trustworthy deployment of AI systems necessitates addressing privacy concerns to prevent unintentional disclosure and discrimination. One significant threat, membership inference (MI) attacks, exploit vulnerabilities in target learning models to determine if a given sample was part of the training set. However, the effectiveness of existing MI attacks is often limited by the number of classes in the dataset or the need for diverse multilevel adversarial features to exploit overfitted models. To enhance MI attack performance, we propose meMIA, a novel framework based on stacked ensemble learning. meMIA integrates embeddings from a neural network (NN) and a long short-term memory (LSTM) model, training a subsequent NN, termed the meta-model, on the concatenated embeddings. This method leverages the complementary strengths of NN and LSTM models; the LSTM captures order differences in confidence scores, while the NN discerns probability distribution differences between member and nonmember samples. We extensively evaluate meMIA on seven benchmark datasets, demonstrating that it surpasses current state-of-the-art MI attacks, achieving accuracy up to 94.6% and near-perfect recall. meMIA's superior performance, especially on datasets with fewer classes, underscores the urgent need for robust defenses against privacy attacks in machine learning, contributing to the safer and more ethical use of AI technologies.
机器学习模型中私人信息的泄露会导致泄密、身份盗窃和未经授权访问个人数据。要确保安全、可靠地部署人工智能系统,就必须解决隐私问题,防止无意泄露和歧视。其中一个重要威胁是成员推理(MI)攻击,它利用目标学习模型中的漏洞来确定给定样本是否是训练集的一部分。然而,现有 MI 攻击的有效性往往受限于数据集中的类别数量,或需要利用过度拟合模型的多样化多层次对抗特征。为了提高 MI 攻击性能,我们提出了基于堆叠集合学习的新型框架 meMIA。meMIA 整合了神经网络(NN)和长短期记忆(LSTM)模型的嵌入,并在连接嵌入上训练后续的 NN(称为元模型)。这种方法利用了 NN 模型和 LSTM 模型的互补优势;LSTM 可捕捉置信度得分的顺序差异,而 NN 则可识别成员样本和非成员样本之间的概率分布差异。我们在七个基准数据集上对meMIA进行了广泛评估,结果表明它超越了当前最先进的MI攻击,准确率高达94.6%,召回率也接近完美。meMIA的卓越性能,尤其是在类别较少的数据集上的性能,凸显了在机器学习中对隐私攻击进行稳健防御的迫切需要,有助于更安全、更合乎道德地使用人工智能技术。