HoneyModels: Machine Learning Honeypots

Ahmed Abdou, Ryan Sheatsley, Yohan Beugin, Tyler J. Shipp, P. Mcdaniel
{"title":"HoneyModels: Machine Learning Honeypots","authors":"Ahmed Abdou, Ryan Sheatsley, Yohan Beugin, Tyler J. Shipp, P. Mcdaniel","doi":"10.1109/MILCOM52596.2021.9652947","DOIUrl":null,"url":null,"abstract":"Machine Learning is becoming a pivotal aspect of many systems today, offering newfound performance on classification and prediction tasks, but this rapid integration also comes with new unforeseen vulnerabilities. To harden these systems the ever-growing field of Adversarial Machine Learning has proposed new attack and defense mechanisms. However, a great asymmetry exists as these defensive methods can only provide security to certain models and lack scalability, computational efficiency, and practicality due to overly restrictive constraints. Moreover, newly introduced attacks can easily bypass defensive strategies by making subtle alterations. In this paper, we study an alternate approach inspired by honeypots to detect adversaries. Our approach yields learned models with an embedded watermark. When an adversary initiates an interaction with our model, attacks are encouraged to add this predetermined watermark stimulating detection of adversarial examples. We show that HoneyModels can reveal 69.5% of adversaries attempting to attack a Neural Network while preserving the original functionality of the model. HoneyModels offer an alternate direction to secure Machine Learning that slightly affects the accuracy while encouraging the creation of watermarked adversarial samples detectable by the HoneyModel but indistinguishable from others for the adversary.","PeriodicalId":187645,"journal":{"name":"MILCOM 2021 - 2021 IEEE Military Communications Conference (MILCOM)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"MILCOM 2021 - 2021 IEEE Military Communications Conference (MILCOM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MILCOM52596.2021.9652947","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Machine Learning is becoming a pivotal aspect of many systems today, offering newfound performance on classification and prediction tasks, but this rapid integration also comes with new unforeseen vulnerabilities. To harden these systems the ever-growing field of Adversarial Machine Learning has proposed new attack and defense mechanisms. However, a great asymmetry exists as these defensive methods can only provide security to certain models and lack scalability, computational efficiency, and practicality due to overly restrictive constraints. Moreover, newly introduced attacks can easily bypass defensive strategies by making subtle alterations. In this paper, we study an alternate approach inspired by honeypots to detect adversaries. Our approach yields learned models with an embedded watermark. When an adversary initiates an interaction with our model, attacks are encouraged to add this predetermined watermark stimulating detection of adversarial examples. We show that HoneyModels can reveal 69.5% of adversaries attempting to attack a Neural Network while preserving the original functionality of the model. HoneyModels offer an alternate direction to secure Machine Learning that slightly affects the accuracy while encouraging the creation of watermarked adversarial samples detectable by the HoneyModel but indistinguishable from others for the adversary.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
HoneyModels:机器学习蜜罐
机器学习正在成为当今许多系统的关键方面,在分类和预测任务上提供了新的性能,但这种快速集成也带来了新的不可预见的漏洞。为了强化这些系统,不断发展的对抗性机器学习领域提出了新的攻击和防御机制。然而,这些防御方法由于过于严格的约束,只能为某些模型提供安全性,缺乏可扩展性、计算效率和实用性,存在很大的不对称性。此外,新引入的攻击可以通过细微的改变轻易绕过防御策略。在本文中,我们研究了一种受蜜罐启发的替代方法来检测对手。我们的方法产生带有嵌入水印的学习模型。当攻击者发起与我们的模型的交互时,攻击者被鼓励添加这个预定的水印来刺激对抗性样本的检测。我们表明,HoneyModels可以在保留模型原始功能的同时,揭示69.5%的攻击者试图攻击神经网络。HoneyModels提供了另一种方法来确保机器学习的准确性,同时鼓励创建被HoneyModel检测到的带水印的对抗样本,但对于对手来说,与其他样本无法区分。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
RF-based Network Inference: Theoretical Foundations Security Threats Analysis of the Unmanned Aerial Vehicle System Using Distributed Ledgers For Command and Control – Concepts and Challenges DerechoDDS: Strongly Consistent Data Distribution for Mission-Critical Applications CUE: A Standalone Testbed for 5G Experimentation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1