Charles Meyers, Mohammad Reza Saleh Sedghpour, Tommy Löfstedt, Erik Elmroth
{"title":"神经网络对抗鲁棒性的成本意识方法","authors":"Charles Meyers, Mohammad Reza Saleh Sedghpour, Tommy Löfstedt, Erik Elmroth","doi":"arxiv-2409.07609","DOIUrl":null,"url":null,"abstract":"Considering the growing prominence of production-level AI and the threat of\nadversarial attacks that can evade a model at run-time, evaluating the\nrobustness of models to these evasion attacks is of critical importance.\nAdditionally, testing model changes likely means deploying the models to (e.g.\na car or a medical imaging device), or a drone to see how it affects\nperformance, making un-tested changes a public problem that reduces development\nspeed, increases cost of development, and makes it difficult (if not\nimpossible) to parse cause from effect. In this work, we used survival analysis\nas a cloud-native, time-efficient and precise method for predicting model\nperformance in the presence of adversarial noise. For neural networks in\nparticular, the relationships between the learning rate, batch size, training\ntime, convergence time, and deployment cost are highly complex, so researchers\ngenerally rely on benchmark datasets to assess the ability of a model to\ngeneralize beyond the training data. To address this, we propose using\naccelerated failure time models to measure the effect of hardware choice, batch\nsize, number of epochs, and test-set accuracy by using adversarial attacks to\ninduce failures on a reference model architecture before deploying the model to\nthe real world. We evaluate several GPU types and use the Tree Parzen Estimator\nto maximize model robustness and minimize model run-time simultaneously. This\nprovides a way to evaluate the model and optimise it in a single step, while\nsimultaneously allowing us to model the effect of model parameters on training\ntime, prediction time, and accuracy. Using this technique, we demonstrate that\nnewer, more-powerful hardware does decrease the training time, but with a\nmonetary and power cost that far outpaces the marginal gains in accuracy.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Cost-Aware Approach to Adversarial Robustness in Neural Networks\",\"authors\":\"Charles Meyers, Mohammad Reza Saleh Sedghpour, Tommy Löfstedt, Erik Elmroth\",\"doi\":\"arxiv-2409.07609\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Considering the growing prominence of production-level AI and the threat of\\nadversarial attacks that can evade a model at run-time, evaluating the\\nrobustness of models to these evasion attacks is of critical importance.\\nAdditionally, testing model changes likely means deploying the models to (e.g.\\na car or a medical imaging device), or a drone to see how it affects\\nperformance, making un-tested changes a public problem that reduces development\\nspeed, increases cost of development, and makes it difficult (if not\\nimpossible) to parse cause from effect. In this work, we used survival analysis\\nas a cloud-native, time-efficient and precise method for predicting model\\nperformance in the presence of adversarial noise. For neural networks in\\nparticular, the relationships between the learning rate, batch size, training\\ntime, convergence time, and deployment cost are highly complex, so researchers\\ngenerally rely on benchmark datasets to assess the ability of a model to\\ngeneralize beyond the training data. To address this, we propose using\\naccelerated failure time models to measure the effect of hardware choice, batch\\nsize, number of epochs, and test-set accuracy by using adversarial attacks to\\ninduce failures on a reference model architecture before deploying the model to\\nthe real world. We evaluate several GPU types and use the Tree Parzen Estimator\\nto maximize model robustness and minimize model run-time simultaneously. This\\nprovides a way to evaluate the model and optimise it in a single step, while\\nsimultaneously allowing us to model the effect of model parameters on training\\ntime, prediction time, and accuracy. Using this technique, we demonstrate that\\nnewer, more-powerful hardware does decrease the training time, but with a\\nmonetary and power cost that far outpaces the marginal gains in accuracy.\",\"PeriodicalId\":501172,\"journal\":{\"name\":\"arXiv - STAT - Applications\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - STAT - Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.07609\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07609","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Cost-Aware Approach to Adversarial Robustness in Neural Networks
Considering the growing prominence of production-level AI and the threat of
adversarial attacks that can evade a model at run-time, evaluating the
robustness of models to these evasion attacks is of critical importance.
Additionally, testing model changes likely means deploying the models to (e.g.
a car or a medical imaging device), or a drone to see how it affects
performance, making un-tested changes a public problem that reduces development
speed, increases cost of development, and makes it difficult (if not
impossible) to parse cause from effect. In this work, we used survival analysis
as a cloud-native, time-efficient and precise method for predicting model
performance in the presence of adversarial noise. For neural networks in
particular, the relationships between the learning rate, batch size, training
time, convergence time, and deployment cost are highly complex, so researchers
generally rely on benchmark datasets to assess the ability of a model to
generalize beyond the training data. To address this, we propose using
accelerated failure time models to measure the effect of hardware choice, batch
size, number of epochs, and test-set accuracy by using adversarial attacks to
induce failures on a reference model architecture before deploying the model to
the real world. We evaluate several GPU types and use the Tree Parzen Estimator
to maximize model robustness and minimize model run-time simultaneously. This
provides a way to evaluate the model and optimise it in a single step, while
simultaneously allowing us to model the effect of model parameters on training
time, prediction time, and accuracy. Using this technique, we demonstrate that
newer, more-powerful hardware does decrease the training time, but with a
monetary and power cost that far outpaces the marginal gains in accuracy.