人工智能分类算法中类不平衡的鲁棒性

IF 2.6 2区 工程技术 Q2 ENGINEERING, INDUSTRIAL Journal of Quality Technology Pub Date : 2021-08-30 DOI:10.1080/00224065.2021.1963200
J. Lian, Laura J. Freeman, Yili Hong, Xinwei Deng
{"title":"人工智能分类算法中类不平衡的鲁棒性","authors":"J. Lian, Laura J. Freeman, Yili Hong, Xinwei Deng","doi":"10.1080/00224065.2021.1963200","DOIUrl":null,"url":null,"abstract":"Abstract Artificial intelligence (AI) algorithms, such as deep learning and XGboost, are used in numerous applications including autonomous driving, manufacturing process optimization and medical diagnostics. The robustness of AI algorithms is of great interest as inaccurate prediction could result in safety concerns and limit the adoption of AI systems. In this paper, we propose a framework based on design of experiments to systematically investigate the robustness of AI classification algorithms. A robust classification algorithm is expected to have high accuracy and low variability under different application scenarios. The robustness can be affected by a wide range of factors such as the imbalance of class labels in the training dataset, the chosen prediction algorithm, the chosen dataset of the application, and a change of distribution in the training and test datasets. To investigate the robustness of AI classification algorithms, we conduct a comprehensive set of mixture experiments to collect prediction performance results. Then statistical analyses are conducted to understand how various factors affect the robustness of AI classification algorithms. We summarize our findings and provide suggestions to practitioners in AI applications.","PeriodicalId":54769,"journal":{"name":"Journal of Quality Technology","volume":"303 1","pages":"505 - 525"},"PeriodicalIF":2.6000,"publicationDate":"2021-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"Robustness with respect to class imbalance in artificial intelligence classification algorithms\",\"authors\":\"J. Lian, Laura J. Freeman, Yili Hong, Xinwei Deng\",\"doi\":\"10.1080/00224065.2021.1963200\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Abstract Artificial intelligence (AI) algorithms, such as deep learning and XGboost, are used in numerous applications including autonomous driving, manufacturing process optimization and medical diagnostics. The robustness of AI algorithms is of great interest as inaccurate prediction could result in safety concerns and limit the adoption of AI systems. In this paper, we propose a framework based on design of experiments to systematically investigate the robustness of AI classification algorithms. A robust classification algorithm is expected to have high accuracy and low variability under different application scenarios. The robustness can be affected by a wide range of factors such as the imbalance of class labels in the training dataset, the chosen prediction algorithm, the chosen dataset of the application, and a change of distribution in the training and test datasets. To investigate the robustness of AI classification algorithms, we conduct a comprehensive set of mixture experiments to collect prediction performance results. Then statistical analyses are conducted to understand how various factors affect the robustness of AI classification algorithms. We summarize our findings and provide suggestions to practitioners in AI applications.\",\"PeriodicalId\":54769,\"journal\":{\"name\":\"Journal of Quality Technology\",\"volume\":\"303 1\",\"pages\":\"505 - 525\"},\"PeriodicalIF\":2.6000,\"publicationDate\":\"2021-08-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Quality Technology\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1080/00224065.2021.1963200\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, INDUSTRIAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Quality Technology","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1080/00224065.2021.1963200","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, INDUSTRIAL","Score":null,"Total":0}
引用次数: 8

摘要

人工智能(AI)算法,如深度学习和XGboost,被用于自动驾驶、制造流程优化和医疗诊断等众多应用中。人工智能算法的鲁棒性非常有趣,因为不准确的预测可能导致安全问题并限制人工智能系统的采用。在本文中,我们提出了一个基于实验设计的框架来系统地研究AI分类算法的鲁棒性。在不同的应用场景下,鲁棒分类算法应具有较高的准确率和较低的可变性。鲁棒性可能受到多种因素的影响,如训练数据集中类标签的不平衡、所选择的预测算法、应用程序所选择的数据集以及训练和测试数据集中分布的变化。为了研究人工智能分类算法的鲁棒性,我们进行了一组全面的混合实验来收集预测性能结果。然后进行统计分析,了解各种因素如何影响人工智能分类算法的鲁棒性。我们总结了我们的发现,并为人工智能应用的从业者提供了建议。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Robustness with respect to class imbalance in artificial intelligence classification algorithms
Abstract Artificial intelligence (AI) algorithms, such as deep learning and XGboost, are used in numerous applications including autonomous driving, manufacturing process optimization and medical diagnostics. The robustness of AI algorithms is of great interest as inaccurate prediction could result in safety concerns and limit the adoption of AI systems. In this paper, we propose a framework based on design of experiments to systematically investigate the robustness of AI classification algorithms. A robust classification algorithm is expected to have high accuracy and low variability under different application scenarios. The robustness can be affected by a wide range of factors such as the imbalance of class labels in the training dataset, the chosen prediction algorithm, the chosen dataset of the application, and a change of distribution in the training and test datasets. To investigate the robustness of AI classification algorithms, we conduct a comprehensive set of mixture experiments to collect prediction performance results. Then statistical analyses are conducted to understand how various factors affect the robustness of AI classification algorithms. We summarize our findings and provide suggestions to practitioners in AI applications.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Journal of Quality Technology
Journal of Quality Technology 管理科学-工程:工业
CiteScore
5.20
自引率
4.00%
发文量
23
审稿时长
>12 weeks
期刊介绍: The objective of Journal of Quality Technology is to contribute to the technical advancement of the field of quality technology by publishing papers that emphasize the practical applicability of new techniques, instructive examples of the operation of existing techniques and results of historical researches. Expository, review, and tutorial papers are also acceptable if they are written in a style suitable for practicing engineers. Sample our Mathematics & Statistics journals, sign in here to start your FREE access for 14 days
期刊最新文献
SpTe2M: An R package for nonparametric modeling and monitoring of spatiotemporal data Spatial modeling and monitoring considering long-range dependence A change-point–based control chart for detecting sparse mean changes in high-dimensional heteroscedastic data Best practices for multi- and mixed-level supersaturated designs Sequential Latin hypercube design for two-layer computer simulators
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1