A new pipeline with ultimate search efficiency for neural architecture search

IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Neural Networks Pub Date : 2025-01-17 DOI:10.1016/j.neunet.2025.107163
Wenbo Liu, Xiaoyun Qiao, Chunyu Zhao, Tao Deng, Fei Yan
{"title":"A new pipeline with ultimate search efficiency for neural architecture search","authors":"Wenbo Liu,&nbsp;Xiaoyun Qiao,&nbsp;Chunyu Zhao,&nbsp;Tao Deng,&nbsp;Fei Yan","doi":"10.1016/j.neunet.2025.107163","DOIUrl":null,"url":null,"abstract":"<div><div>We present a novel neural architecture search pipeline designed to enhance search efficiency through optimized data and algorithms. Leveraging dataset distillation techniques, our pipeline condenses large-scale target datasets into more streamlined proxy datasets, effectively reducing the computational overhead associated with identifying optimal neural architectures. To accommodate diverse approaches to synthetic dataset utilization, our pipeline comprises two distinct schemes. Scheme 1 involves constructing rich data from various Bases <span><math><mrow><mo>|</mo><mi>B</mi><mo>|</mo></mrow></math></span>, while Scheme 2 focuses on establishing high-quality relationship mappings within the data. Models generated through Scheme 1 exhibit outstanding scalability, demonstrating superior performance when transferred to larger, more complex tasks. Despite utilizing fewer data, Scheme 2 maintains performance levels without degradation on the source dataset. Furthermore, our research extends to the inherent challenges present in DARTS-derived algorithms, particularly in the selection of candidate operations based on architectural parameters. We identify architectural parameter disparities across different edges, highlighting the occurrence of “Selection Errors” during the model generation process, and propose an enhanced search algorithm. Our proposed algorithm comprises three components—attention, regularization, and normalization—aiding in the rapid identification of high-quality models using data generated from proxy datasets. Experimental results demonstrate a significant reduction in search time, with high-quality models generated in as little as two minutes using our proposed pipeline. Through comprehensive experimentation, we meticulously validate the efficacy of both schemes and algorithms.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"185 ","pages":"Article 107163"},"PeriodicalIF":6.3000,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025000425","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

We present a novel neural architecture search pipeline designed to enhance search efficiency through optimized data and algorithms. Leveraging dataset distillation techniques, our pipeline condenses large-scale target datasets into more streamlined proxy datasets, effectively reducing the computational overhead associated with identifying optimal neural architectures. To accommodate diverse approaches to synthetic dataset utilization, our pipeline comprises two distinct schemes. Scheme 1 involves constructing rich data from various Bases |B|, while Scheme 2 focuses on establishing high-quality relationship mappings within the data. Models generated through Scheme 1 exhibit outstanding scalability, demonstrating superior performance when transferred to larger, more complex tasks. Despite utilizing fewer data, Scheme 2 maintains performance levels without degradation on the source dataset. Furthermore, our research extends to the inherent challenges present in DARTS-derived algorithms, particularly in the selection of candidate operations based on architectural parameters. We identify architectural parameter disparities across different edges, highlighting the occurrence of “Selection Errors” during the model generation process, and propose an enhanced search algorithm. Our proposed algorithm comprises three components—attention, regularization, and normalization—aiding in the rapid identification of high-quality models using data generated from proxy datasets. Experimental results demonstrate a significant reduction in search time, with high-quality models generated in as little as two minutes using our proposed pipeline. Through comprehensive experimentation, we meticulously validate the efficacy of both schemes and algorithms.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
一种新的具有最高搜索效率的神经结构搜索管道。
我们提出了一种新的神经结构搜索管道,旨在通过优化数据和算法来提高搜索效率。利用数据集蒸馏技术,我们的管道将大规模目标数据集压缩成更精简的代理数据集,有效地减少了与识别最佳神经架构相关的计算开销。为了适应合成数据集利用的不同方法,我们的管道包括两个不同的方案。方案1涉及从各种base中构建丰富的数据,而方案2侧重于在数据中建立高质量的关系映射。通过Scheme 1生成的模型具有出色的可伸缩性,在转移到更大、更复杂的任务时表现出卓越的性能。尽管使用更少的数据,Scheme 2在源数据集上保持了性能水平,而不会降低性能。此外,我们的研究扩展到darts衍生算法中存在的固有挑战,特别是在基于架构参数的候选操作选择方面。我们识别了不同边缘的建筑参数差异,突出了模型生成过程中“选择错误”的发生,并提出了一种增强的搜索算法。我们提出的算法包括三个组成部分——注意、正则化和规范化——帮助使用代理数据集生成的数据快速识别高质量的模型。实验结果表明,搜索时间显著减少,使用我们提出的管道在短短两分钟内生成高质量的模型。通过全面的实验,我们仔细验证了方案和算法的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Neural Networks
Neural Networks 工程技术-计算机:人工智能
CiteScore
13.90
自引率
7.70%
发文量
425
审稿时长
67 days
期刊介绍: Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.
期刊最新文献
Beyond local aggregation: Global graph contrastive learning for multi-view fusion. SD2-ReID: A semantic-stylistic decoupled distillation framework for robust multi-modal object re-identification. TransUTD: Underwater cross-domain collaborative spatial-temporal transformer detector. Adversarial discriminant attack on text-to-image diffusion models. Enhancing out-of-distribution detection with bilateral distribution score.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1