AutoML: A systematic review on automated machine learning with neural architecture search

Imrus Salehin , Md. Shamiul Islam , Pritom Saha , S.M. Noman , Azra Tuni , Md. Mehedi Hasan , Md. Abu Baten
{"title":"AutoML: A systematic review on automated machine learning with neural architecture search","authors":"Imrus Salehin ,&nbsp;Md. Shamiul Islam ,&nbsp;Pritom Saha ,&nbsp;S.M. Noman ,&nbsp;Azra Tuni ,&nbsp;Md. Mehedi Hasan ,&nbsp;Md. Abu Baten","doi":"10.1016/j.jiixd.2023.10.002","DOIUrl":null,"url":null,"abstract":"<div><p>AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 1","pages":"Pages 52-81"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715923000604/pdfft?md5=a79f7fb3cdab55edd3b7838063f99f50&pid=1-s2.0-S2949715923000604-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Information and Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949715923000604","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
AutoML:利用神经架构搜索自动机器学习的系统综述
AutoML(自动化机器学习)是一个新兴领域,旨在实现机器学习模型构建过程的自动化。AutoML 的出现是为了尽可能自动化重复机器学习过程中出现的低效工作,从而提高生产率和效率。特别是,从数据预处理到算法选择和调整,模型开发人员在这一过程中的干预降到最低,从而有效开发出高质量模型的技术已经研究了很长时间。在这项语义回顾研究中,我们总结了 AutoML 方法的数据处理要求,并提供了详细的解释。我们更加重视神经架构搜索(NAS),因为它是目前 AutoML 领域非常热门的子课题。NAS 方法使用机器学习算法在大量可能的架构中进行搜索,找出在给定任务中表现最佳的架构。我们总结了具有代表性的 NAS 算法在 CIFAR-10、CIFAR-100、ImageNet 和知名基准数据集上取得的性能。此外,我们还深入探讨了 NAS 方法中几个值得关注的研究方向,包括单/两阶段 NAS、单次 NAS 和联合超参数与架构优化。我们讨论了 NAS 的搜索空间大小和复杂性如何因所解决的具体问题而异。最后,我们探讨了当前 AutoML 方法中的几个开放问题(SOTA 问题),这些问题值得在未来的研究中进一步探讨。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Editorial Board Editorial Board Secure performance comparison for NOMA: Reconfigurable intelligent surface or amplify-and-forward relay? Editorial Board Structural knowledge-driven meta-learning for task offloading in vehicular networks with integrated communications, sensing and computing
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1