首页 > 最新文献

Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security最新文献

英文 中文
Session details: Keynote Address 会议详情:主题演讲
B. Biggio
{"title":"Session details: Keynote Address","authors":"B. Biggio","doi":"10.1145/3252885","DOIUrl":"https://doi.org/10.1145/3252885","url":null,"abstract":"","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132508260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mitigating Poisoning Attacks on Machine Learning Models: A Data Provenance Based Approach 减轻对机器学习模型的中毒攻击:一种基于数据来源的方法
Pub Date : 2017-11-03 DOI: 10.1145/3128572.3140450
N. Baracaldo, Bryant Chen, Heiko Ludwig, Jaehoon Amir Safavi
The use of machine learning models has become ubiquitous. Their predictions are used to make decisions about healthcare, security, investments and many other critical applications. Given this pervasiveness, it is not surprising that adversaries have an incentive to manipulate machine learning models to their advantage. One way of manipulating a model is through a poisoning or causative attack in which the adversary feeds carefully crafted poisonous data points into the training set. Taking advantage of recently developed tamper-free provenance frameworks, we present a methodology that uses contextual information about the origin and transformation of data points in the training set to identify poisonous data, thereby enabling online and regularly re-trained machine learning applications to consume data sources in potentially adversarial environments. To the best of our knowledge, this is the first approach to incorporate provenance information as part of a filtering algorithm to detect causative attacks. We present two variations of the methodology - one tailored to partially trusted data sets and the other to fully untrusted data sets. Finally, we evaluate our methodology against existing methods to detect poison data and show an improvement in the detection rate.
机器学习模型的使用已经变得无处不在。他们的预测用于制定有关医疗保健、安全、投资和许多其他关键应用程序的决策。鉴于这种普遍性,对手有动机操纵机器学习模型以获得优势也就不足为奇了。操纵模型的一种方法是通过中毒攻击或因果攻击,在这种攻击中,对手将精心制作的有毒数据点输入训练集中。利用最近开发的无篡改来源框架,我们提出了一种方法,该方法使用有关训练集中数据点的起源和转换的上下文信息来识别有毒数据,从而使在线和定期重新训练的机器学习应用程序能够在潜在的敌对环境中使用数据源。据我们所知,这是第一个将来源信息作为过滤算法的一部分来检测病因攻击的方法。我们提出了该方法的两种变体-一种针对部分可信数据集,另一种针对完全不可信数据集。最后,我们评估了我们的方法对现有的方法来检测毒物数据,并显示在检出率的改进。
{"title":"Mitigating Poisoning Attacks on Machine Learning Models: A Data Provenance Based Approach","authors":"N. Baracaldo, Bryant Chen, Heiko Ludwig, Jaehoon Amir Safavi","doi":"10.1145/3128572.3140450","DOIUrl":"https://doi.org/10.1145/3128572.3140450","url":null,"abstract":"The use of machine learning models has become ubiquitous. Their predictions are used to make decisions about healthcare, security, investments and many other critical applications. Given this pervasiveness, it is not surprising that adversaries have an incentive to manipulate machine learning models to their advantage. One way of manipulating a model is through a poisoning or causative attack in which the adversary feeds carefully crafted poisonous data points into the training set. Taking advantage of recently developed tamper-free provenance frameworks, we present a methodology that uses contextual information about the origin and transformation of data points in the training set to identify poisonous data, thereby enabling online and regularly re-trained machine learning applications to consume data sources in potentially adversarial environments. To the best of our knowledge, this is the first approach to incorporate provenance information as part of a filtering algorithm to detect causative attacks. We present two variations of the methodology - one tailored to partially trusted data sets and the other to fully untrusted data sets. Finally, we evaluate our methodology against existing methods to detect poison data and show an improvement in the detection rate.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125940975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 80
In (Cyber)Space Bots Can Hear You Speak: Breaking Audio CAPTCHAs Using OTS Speech Recognition 在(网络)空间机器人可以听到你说话:打破音频验证码使用OTS语音识别
Pub Date : 2017-11-03 DOI: 10.1145/3128572.3140443
Saumya Solanki, Gautam Krishnan, Varshini Sampath, Jason Polakis
Captchas have become almost ubiquitous as they are commonly deployed by websites as part of their defenses against fraudsters. However visual captchas pose a considerable obstacle to certain groups of users, such as the visually impaired, and that has necessitated the inclusion of more accessible captcha schemes. As a result, many captcha services also offer audio challenges as an alternative. In this paper we conduct an extensive exploration of the audio captcha ecosystem, and present effective low-cost attacks against the audio challenges offered by seven major captcha services. Motivated by the recent advancements in deep learning, we demonstrate how off-the-shelf (OTS) speech recognition services can be misused by attackers for trivially bypassing the most popular audio captchas. Our experimental evaluation highlights the effectiveness of our approach as our AudioBreaker system is able to break all captcha schemes, achieving accuracies of up to 98.3% against Google's ReCaptcha. The broader implications of our study are twofold. First, we find that the wide availability of advanced speech recognition services has severely lowered the technical capabilities required by fraudsters for deploying effective attacks, as there is no longer a need to build sophisticated custom classifiers. Second, we find that the availability of audio captchas poses a significant risk to services, as our attacks against ReCaptcha's audio challenges are 13.1%-27.5% more accurate than state-of-the-art attacks against the corresponding image-based challenges. Overall, we argue that it is necessary to explore alternative captcha designs that fulfill the accessibility properties of audio captchas without undermining the security offered by their visual counterparts.
验证码已经变得几乎无处不在,因为它们通常被网站部署,作为防御欺诈者的一部分。然而,视觉验证码对某些用户群体(如视障人士)构成了相当大的障碍,这就需要包含更容易访问的验证码方案。因此,许多验证码服务也提供音频挑战作为替代方案。在本文中,我们对音频验证码生态系统进行了广泛的探索,并提出了针对七种主要验证码服务提供的音频挑战的有效低成本攻击。在深度学习最新进展的推动下,我们展示了现成的(OTS)语音识别服务如何被攻击者滥用,以轻松绕过最流行的音频验证码。我们的实验评估强调了我们方法的有效性,因为我们的AudioBreaker系统能够打破所有验证码方案,对b谷歌的验证码实现高达98.3%的准确率。我们的研究有两个更广泛的含义。首先,我们发现先进语音识别服务的广泛可用性严重降低了欺诈者部署有效攻击所需的技术能力,因为不再需要构建复杂的自定义分类器。其次,我们发现音频验证码的可用性对服务构成了重大风险,因为我们针对音频验证码的攻击比针对相应基于图像的挑战的最先进攻击的准确率高13.1%-27.5%。总的来说,我们认为有必要探索替代验证码设计,以满足音频验证码的可访问性属性,而不会破坏其视觉对应物提供的安全性。
{"title":"In (Cyber)Space Bots Can Hear You Speak: Breaking Audio CAPTCHAs Using OTS Speech Recognition","authors":"Saumya Solanki, Gautam Krishnan, Varshini Sampath, Jason Polakis","doi":"10.1145/3128572.3140443","DOIUrl":"https://doi.org/10.1145/3128572.3140443","url":null,"abstract":"Captchas have become almost ubiquitous as they are commonly deployed by websites as part of their defenses against fraudsters. However visual captchas pose a considerable obstacle to certain groups of users, such as the visually impaired, and that has necessitated the inclusion of more accessible captcha schemes. As a result, many captcha services also offer audio challenges as an alternative. In this paper we conduct an extensive exploration of the audio captcha ecosystem, and present effective low-cost attacks against the audio challenges offered by seven major captcha services. Motivated by the recent advancements in deep learning, we demonstrate how off-the-shelf (OTS) speech recognition services can be misused by attackers for trivially bypassing the most popular audio captchas. Our experimental evaluation highlights the effectiveness of our approach as our AudioBreaker system is able to break all captcha schemes, achieving accuracies of up to 98.3% against Google's ReCaptcha. The broader implications of our study are twofold. First, we find that the wide availability of advanced speech recognition services has severely lowered the technical capabilities required by fraudsters for deploying effective attacks, as there is no longer a need to build sophisticated custom classifiers. Second, we find that the availability of audio captchas poses a significant risk to services, as our attacks against ReCaptcha's audio challenges are 13.1%-27.5% more accurate than state-of-the-art attacks against the corresponding image-based challenges. Overall, we argue that it is necessary to explore alternative captcha designs that fulfill the accessibility properties of audio captchas without undermining the security offered by their visual counterparts.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124032767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Session details: Deep Learning 会议详情:深度学习
David Mandell Freeman
{"title":"Session details: Deep Learning","authors":"David Mandell Freeman","doi":"10.1145/3252886","DOIUrl":"https://doi.org/10.1145/3252886","url":null,"abstract":"","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114055398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Practical Machine Learning for Cloud Intrusion Detection: Challenges and the Way Forward 云入侵检测的实用机器学习:挑战和前进的方向
Pub Date : 2017-09-20 DOI: 10.1145/3128572.3140445
R. Kumar, Andrew W. Wicker, Matt Swann
Operationalizing machine learning based security detections is extremely challenging, especially in a continuously evolving cloud environment. Conventional anomaly detection does not produce satisfactory results for analysts that are investigating security incidents in the cloud. Model evaluation alone presents its own set of problems due to a lack of benchmark datasets. When deploying these detections, we must deal with model compliance, localization, and data silo issues, among many others. We pose the problem of "attack disruption" as a way forward in the security data science space. In this paper, we describe the framework, challenges, and open questions surrounding the successful operationalization of machine learning based security detections in a cloud environment and provide some insights on how we have addressed them.
实施基于机器学习的安全检测极具挑战性,特别是在不断发展的云环境中。传统的异常检测不能为调查云中的安全事件的分析人员产生令人满意的结果。由于缺乏基准数据集,模型评估本身就存在一系列问题。在部署这些检测时,我们必须处理模型遵从性、本地化和数据竖井等问题。我们将“攻击中断”问题作为安全数据科学领域的前进方向。在本文中,我们描述了在云环境中成功实施基于机器学习的安全检测的框架、挑战和开放问题,并就我们如何解决这些问题提供了一些见解。
{"title":"Practical Machine Learning for Cloud Intrusion Detection: Challenges and the Way Forward","authors":"R. Kumar, Andrew W. Wicker, Matt Swann","doi":"10.1145/3128572.3140445","DOIUrl":"https://doi.org/10.1145/3128572.3140445","url":null,"abstract":"Operationalizing machine learning based security detections is extremely challenging, especially in a continuously evolving cloud environment. Conventional anomaly detection does not produce satisfactory results for analysts that are investigating security incidents in the cloud. Model evaluation alone presents its own set of problems due to a lack of benchmark datasets. When deploying these detections, we must deal with model compliance, localization, and data silo issues, among many others. We pose the problem of \"attack disruption\" as a way forward in the security data science space. In this paper, we describe the framework, challenges, and open questions surrounding the successful operationalization of machine learning based security detections in a cloud environment and provide some insights on how we have addressed them.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121848909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Learning the PE Header, Malware Detection with Minimal Domain Knowledge 学习PE头,恶意软件检测与最小的领域知识
Pub Date : 2017-09-05 DOI: 10.1145/3128572.3140442
Edward Raff, Jared Sylvester, Charles K. Nicholas
Many efforts have been made to use various forms of domain knowledge in malware detection. Currently there exist two common approaches to malware detection without domain knowledge, namely byte n-grams and strings. In this work we explore the feasibility of applying neural networks to malware detection and feature learning. We do this by restricting ourselves to a minimal amount of domain knowledge in order to extract a portion of the Portable Executable (PE) header. By doing this we show that neural networks can learn from raw bytes without explicit feature construction, and perform even better than a domain knowledge approach that parses the PE header into explicit features.
在恶意软件检测中使用各种形式的领域知识已经做了很多努力。目前存在两种不需要领域知识的恶意软件检测方法,即字节n-图和字符串。在这项工作中,我们探索了将神经网络应用于恶意软件检测和特征学习的可行性。为了提取可移植可执行文件(PE)头文件的一部分,我们将自己限制在最小的领域知识范围内。通过这样做,我们表明神经网络可以在没有显式特征构建的情况下从原始字节中学习,并且比将PE头解析为显式特征的领域知识方法表现得更好。
{"title":"Learning the PE Header, Malware Detection with Minimal Domain Knowledge","authors":"Edward Raff, Jared Sylvester, Charles K. Nicholas","doi":"10.1145/3128572.3140442","DOIUrl":"https://doi.org/10.1145/3128572.3140442","url":null,"abstract":"Many efforts have been made to use various forms of domain knowledge in malware detection. Currently there exist two common approaches to malware detection without domain knowledge, namely byte n-grams and strings. In this work we explore the feasibility of applying neural networks to malware detection and feature learning. We do this by restricting ourselves to a minimal amount of domain knowledge in order to extract a portion of the Portable Executable (PE) header. By doing this we show that neural networks can learn from raw bytes without explicit feature construction, and perform even better than a domain knowledge approach that parses the PE header into explicit features.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126317358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 103
Towards Poisoning of Deep Learning Algorithms with Back-gradient Optimization 基于反向梯度优化的深度学习算法中毒研究
Pub Date : 2017-08-29 DOI: 10.1145/3128572.3140451
Luis Muñoz-González, B. Biggio, Ambra Demontis, Andrea Paudice, Vasin Wongrassamee, Emil C. Lupu, F. Roli
A number of online services nowadays rely upon machine learning to extract valuable information from data collected in the wild. This exposes learning algorithms to the threat of data poisoning, i.e., a coordinate attack in which a fraction of the training data is controlled by the attacker and manipulated to subvert the learning process. To date, these attacks have been devised only against a limited class of binary learning algorithms, due to the inherent complexity of the gradient-based procedure used to optimize the poisoning points (a.k.a. adversarial training examples). In this work, we first extend the definition of poisoning attacks to multiclass problems. We then propose a novel poisoning algorithm based on the idea of back-gradient optimization, i.e., to compute the gradient of interest through automatic differentiation, while also reversing the learning procedure to drastically reduce the attack complexity. Compared to current poisoning strategies, our approach is able to target a wider class of learning algorithms, trained with gradient-based procedures, including neural networks and deep learning architectures. We empirically evaluate its effectiveness on several application examples, including spam filtering, malware detection, and handwritten digit recognition. We finally show that, similarly to adversarial test examples, adversarial training examples can also be transferred across different learning algorithms.
如今,许多在线服务都依靠机器学习从野外收集的数据中提取有价值的信息。这使得学习算法面临数据中毒的威胁,即一种坐标攻击,其中一小部分训练数据被攻击者控制并被操纵以破坏学习过程。到目前为止,由于用于优化中毒点(也称为对抗性训练示例)的基于梯度的程序的固有复杂性,这些攻击仅针对有限类别的二进制学习算法设计。在这项工作中,我们首先将中毒攻击的定义扩展到多类问题。然后,我们提出了一种基于反向梯度优化思想的新型中毒算法,即通过自动微分计算感兴趣的梯度,同时还反转学习过程以大幅降低攻击复杂度。与目前的中毒策略相比,我们的方法能够针对更广泛的学习算法,使用基于梯度的程序进行训练,包括神经网络和深度学习架构。我们在几个应用实例中对其有效性进行了实证评估,包括垃圾邮件过滤、恶意软件检测和手写数字识别。我们最后表明,与对抗性测试样例类似,对抗性训练样例也可以在不同的学习算法之间转移。
{"title":"Towards Poisoning of Deep Learning Algorithms with Back-gradient Optimization","authors":"Luis Muñoz-González, B. Biggio, Ambra Demontis, Andrea Paudice, Vasin Wongrassamee, Emil C. Lupu, F. Roli","doi":"10.1145/3128572.3140451","DOIUrl":"https://doi.org/10.1145/3128572.3140451","url":null,"abstract":"A number of online services nowadays rely upon machine learning to extract valuable information from data collected in the wild. This exposes learning algorithms to the threat of data poisoning, i.e., a coordinate attack in which a fraction of the training data is controlled by the attacker and manipulated to subvert the learning process. To date, these attacks have been devised only against a limited class of binary learning algorithms, due to the inherent complexity of the gradient-based procedure used to optimize the poisoning points (a.k.a. adversarial training examples). In this work, we first extend the definition of poisoning attacks to multiclass problems. We then propose a novel poisoning algorithm based on the idea of back-gradient optimization, i.e., to compute the gradient of interest through automatic differentiation, while also reversing the learning procedure to drastically reduce the attack complexity. Compared to current poisoning strategies, our approach is able to target a wider class of learning algorithms, trained with gradient-based procedures, including neural networks and deep learning architectures. We empirically evaluate its effectiveness on several application examples, including spam filtering, malware detection, and handwritten digit recognition. We finally show that, similarly to adversarial test examples, adversarial training examples can also be transferred across different learning algorithms.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130677225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 484
ZOO: Zeroth Order Optimization Based Black-box Attacks to Deep Neural Networks without Training Substitute Models 基于零阶优化的无训练替代模型的深度神经网络黑盒攻击
Pub Date : 2017-08-14 DOI: 10.1145/3128572.3140448
Pin-Yu Chen, Huan Zhang, Yash Sharma, Jinfeng Yi, Cho-Jui Hsieh
Deep neural networks (DNNs) are one of the most prominent technologies of our time, as they achieve state-of-the-art performance in many machine learning tasks, including but not limited to image classification, text mining, and speech processing. However, recent research on DNNs has indicated ever-increasing concern on the robustness to adversarial examples, especially for security-critical tasks such as traffic sign identification for autonomous driving. Studies have unveiled the vulnerability of a well-trained DNN by demonstrating the ability of generating barely noticeable (to both human and machines) adversarial images that lead to misclassification. Furthermore, researchers have shown that these adversarial images are highly transferable by simply training and attacking a substitute model built upon the target model, known as a black-box attack to DNNs. Similar to the setting of training substitute models, in this paper we propose an effective black-box attack that also only has access to the input (images) and the output (confidence scores) of a targeted DNN. However, different from leveraging attack transferability from substitute models, we propose zeroth order optimization (ZOO) based attacks to directly estimate the gradients of the targeted DNN for generating adversarial examples. We use zeroth order stochastic coordinate descent along with dimension reduction, hierarchical attack and importance sampling techniques to efficiently attack black-box models. By exploiting zeroth order optimization, improved attacks to the targeted DNN can be accomplished, sparing the need for training substitute models and avoiding the loss in attack transferability. Experimental results on MNIST, CIFAR10 and ImageNet show that the proposed ZOO attack is as effective as the state-of-the-art white-box attack (e.g., Carlini and Wagner's attack) and significantly outperforms existing black-box attacks via substitute models.
深度神经网络(dnn)是我们这个时代最突出的技术之一,因为它们在许多机器学习任务中实现了最先进的性能,包括但不限于图像分类、文本挖掘和语音处理。然而,最近对深度神经网络的研究表明,人们越来越关注对对抗示例的鲁棒性,特别是对于安全关键任务,如自动驾驶的交通标志识别。研究揭示了训练有素的深度神经网络的脆弱性,展示了生成(对人类和机器来说)几乎不明显的对抗图像的能力,这些图像会导致错误分类。此外,研究人员已经证明,通过简单地训练和攻击建立在目标模型之上的替代模型,这些对抗性图像具有高度可转移性,这被称为对dnn的黑盒攻击。与训练替代模型的设置类似,在本文中,我们提出了一种有效的黑盒攻击,该攻击也只能访问目标DNN的输入(图像)和输出(置信度分数)。然而,与利用替代模型的攻击可转移性不同,我们提出了基于零阶优化(ZOO)的攻击来直接估计目标DNN的梯度以生成对抗性示例。采用零阶随机坐标下降、降维、分层攻击和重要采样技术对黑盒模型进行有效攻击。通过利用零阶优化,可以完成对目标DNN的改进攻击,省去了训练替代模型的需要,避免了攻击可转移性的损失。在MNIST, CIFAR10和ImageNet上的实验结果表明,所提出的ZOO攻击与最先进的白盒攻击(例如Carlini和Wagner的攻击)一样有效,并且通过替代模型显著优于现有的黑盒攻击。
{"title":"ZOO: Zeroth Order Optimization Based Black-box Attacks to Deep Neural Networks without Training Substitute Models","authors":"Pin-Yu Chen, Huan Zhang, Yash Sharma, Jinfeng Yi, Cho-Jui Hsieh","doi":"10.1145/3128572.3140448","DOIUrl":"https://doi.org/10.1145/3128572.3140448","url":null,"abstract":"Deep neural networks (DNNs) are one of the most prominent technologies of our time, as they achieve state-of-the-art performance in many machine learning tasks, including but not limited to image classification, text mining, and speech processing. However, recent research on DNNs has indicated ever-increasing concern on the robustness to adversarial examples, especially for security-critical tasks such as traffic sign identification for autonomous driving. Studies have unveiled the vulnerability of a well-trained DNN by demonstrating the ability of generating barely noticeable (to both human and machines) adversarial images that lead to misclassification. Furthermore, researchers have shown that these adversarial images are highly transferable by simply training and attacking a substitute model built upon the target model, known as a black-box attack to DNNs. Similar to the setting of training substitute models, in this paper we propose an effective black-box attack that also only has access to the input (images) and the output (confidence scores) of a targeted DNN. However, different from leveraging attack transferability from substitute models, we propose zeroth order optimization (ZOO) based attacks to directly estimate the gradients of the targeted DNN for generating adversarial examples. We use zeroth order stochastic coordinate descent along with dimension reduction, hierarchical attack and importance sampling techniques to efficiently attack black-box models. By exploiting zeroth order optimization, improved attacks to the targeted DNN can be accomplished, sparing the need for training substitute models and avoiding the loss in attack transferability. Experimental results on MNIST, CIFAR10 and ImageNet show that the proposed ZOO attack is as effective as the state-of-the-art white-box attack (e.g., Carlini and Wagner's attack) and significantly outperforms existing black-box attacks via substitute models.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131098403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1445
Efficient Defenses Against Adversarial Attacks 有效防御对抗性攻击
Pub Date : 2017-07-21 DOI: 10.1145/3128572.3140449
Valentina Zantedeschi, Maria-Irina Nicolae, Ambrish Rawat
Following the recent adoption of deep neural networks (DNN) accross a wide range of applications, adversarial attacks against these models have proven to be an indisputable threat. Adversarial samples are crafted with a deliberate intention of undermining a system. In the case of DNNs, the lack of better understanding of their working has prevented the development of efficient defenses. In this paper, we propose a new defense method based on practical observations which is easy to integrate into models and performs better than state-of-the-art defenses. Our proposed solution is meant to reinforce the structure of a DNN, making its prediction more stable and less likely to be fooled by adversarial samples. We conduct an extensive experimental study proving the efficiency of our method against multiple attacks, comparing it to numerous defenses, both in white-box and black-box setups. Additionally, the implementation of our method brings almost no overhead to the training procedure, while maintaining the prediction performance of the original model on clean samples.
随着深度神经网络(DNN)在广泛应用中的应用,对这些模型的对抗性攻击已被证明是一个无可争议的威胁。对抗性样本是有意破坏系统的。在深层神经网络的情况下,缺乏对其工作的更好理解阻碍了有效防御的发展。在本文中,我们提出了一种新的基于实际观测的防御方法,该方法易于集成到模型中,并且比现有的防御方法性能更好。我们提出的解决方案旨在加强深度神经网络的结构,使其预测更稳定,更不容易被对抗性样本欺骗。我们进行了广泛的实验研究,证明了我们的方法对抗多种攻击的效率,并将其与白盒和黑盒设置中的众多防御进行了比较。此外,我们的方法的实现几乎没有给训练过程带来任何开销,同时在干净样本上保持原始模型的预测性能。
{"title":"Efficient Defenses Against Adversarial Attacks","authors":"Valentina Zantedeschi, Maria-Irina Nicolae, Ambrish Rawat","doi":"10.1145/3128572.3140449","DOIUrl":"https://doi.org/10.1145/3128572.3140449","url":null,"abstract":"Following the recent adoption of deep neural networks (DNN) accross a wide range of applications, adversarial attacks against these models have proven to be an indisputable threat. Adversarial samples are crafted with a deliberate intention of undermining a system. In the case of DNNs, the lack of better understanding of their working has prevented the development of efficient defenses. In this paper, we propose a new defense method based on practical observations which is easy to integrate into models and performs better than state-of-the-art defenses. Our proposed solution is meant to reinforce the structure of a DNN, making its prediction more stable and less likely to be fooled by adversarial samples. We conduct an extensive experimental study proving the efficiency of our method against multiple attacks, comparing it to numerous defenses, both in white-box and black-box setups. Additionally, the implementation of our method brings almost no overhead to the training procedure, while maintaining the prediction performance of the original model on clean samples.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126780731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 262
Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods 不容易检测对抗性示例:绕过十种检测方法
Pub Date : 2017-05-20 DOI: 10.1145/3128572.3140444
Nicholas Carlini, D. Wagner
Neural networks are known to be vulnerable to adversarial examples: inputs that are close to natural inputs but classified incorrectly. In order to better understand the space of adversarial examples, we survey ten recent proposals that are designed for detection and compare their efficacy. We show that all can be defeated by constructing new loss functions. We conclude that adversarial examples are significantly harder to detect than previously appreciated, and the properties believed to be intrinsic to adversarial examples are in fact not. Finally, we propose several simple guidelines for evaluating future proposed defenses.
众所周知,神经网络很容易受到对抗性示例的影响:与自然输入接近但分类错误的输入。为了更好地理解对抗性示例的空间,我们调查了最近设计用于检测的十个建议,并比较了它们的有效性。我们证明了这一切都可以通过构造新的损失函数来克服。我们得出的结论是,对抗性示例比以前所认识到的更难检测,并且被认为是对抗性示例固有的属性实际上并非如此。最后,我们提出了几个简单的准则来评估未来提议的防御。
{"title":"Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods","authors":"Nicholas Carlini, D. Wagner","doi":"10.1145/3128572.3140444","DOIUrl":"https://doi.org/10.1145/3128572.3140444","url":null,"abstract":"Neural networks are known to be vulnerable to adversarial examples: inputs that are close to natural inputs but classified incorrectly. In order to better understand the space of adversarial examples, we survey ten recent proposals that are designed for detection and compare their efficacy. We show that all can be defeated by constructing new loss functions. We conclude that adversarial examples are significantly harder to detect than previously appreciated, and the properties believed to be intrinsic to adversarial examples are in fact not. Finally, we propose several simple guidelines for evaluating future proposed defenses.","PeriodicalId":318259,"journal":{"name":"Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131860889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1622
期刊
Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1