{"title":"SpotOn: A Gradient-based Targeted Data Poisoning Attack on Deep Neural Networks","authors":"Yash Khare, Kumud Lakara, Sparsh Mittal, Arvind Kaushik, Rekha Singhal","doi":"10.1109/ISQED57927.2023.10129311","DOIUrl":null,"url":null,"abstract":"Deep neural networks (DNNs) are vulnerable to adversarial inputs, which are created by adding minor perturbations to the genuine inputs. Previous gradient-based adversarial attacks, such as the \"fast gradient sign method\" (FGSM), add an equal amount (say ϵ) of noise to all the pixels of an image. This degrades image quality significantly, such that a human validator can easily detect the resultant adversarial samples. We propose a novel gradient-based adversarial attack technique named SpotOn, which seeks to maintain the quality of adversarial images high. We first identify an image’s region of importance (ROI) using Grad-CAM. SpotOn has three variants. Two variants of SpotOn attack only the ROI, whereas the third variant adds an epsilon (ϵ) amount of noise to the ROI and a much smaller amount of noise (say ϵ/3) to the remaining image. On Caltech101 dataset, compared to FGSM, SpotOn achieves comparable degradation in CNN accuracy while maintaining much higher image quality. For example, for ϵ = 0.1, FGSM degrades VGG19 accuracy from 92% to 8% and leads to an SSIM value of 0.48 by attacking all pixels in an image. By contrast, SpotOn-VariableNoise attacks only 34.8% of the pixels in the image; degrades accuracy to 10.5% and maintains an SSIM value of 0.78. This makes SpotOn an effective data-poisoning attack technique. The code is available from https://github.com/CandleLabAI/SpotOn-AttackOnDNNs.","PeriodicalId":315053,"journal":{"name":"2023 24th International Symposium on Quality Electronic Design (ISQED)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 24th International Symposium on Quality Electronic Design (ISQED)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISQED57927.2023.10129311","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Deep neural networks (DNNs) are vulnerable to adversarial inputs, which are created by adding minor perturbations to the genuine inputs. Previous gradient-based adversarial attacks, such as the "fast gradient sign method" (FGSM), add an equal amount (say ϵ) of noise to all the pixels of an image. This degrades image quality significantly, such that a human validator can easily detect the resultant adversarial samples. We propose a novel gradient-based adversarial attack technique named SpotOn, which seeks to maintain the quality of adversarial images high. We first identify an image’s region of importance (ROI) using Grad-CAM. SpotOn has three variants. Two variants of SpotOn attack only the ROI, whereas the third variant adds an epsilon (ϵ) amount of noise to the ROI and a much smaller amount of noise (say ϵ/3) to the remaining image. On Caltech101 dataset, compared to FGSM, SpotOn achieves comparable degradation in CNN accuracy while maintaining much higher image quality. For example, for ϵ = 0.1, FGSM degrades VGG19 accuracy from 92% to 8% and leads to an SSIM value of 0.48 by attacking all pixels in an image. By contrast, SpotOn-VariableNoise attacks only 34.8% of the pixels in the image; degrades accuracy to 10.5% and maintains an SSIM value of 0.78. This makes SpotOn an effective data-poisoning attack technique. The code is available from https://github.com/CandleLabAI/SpotOn-AttackOnDNNs.