Xuan Rao, Songyi Xiao, Jiaxin Li, Qiuye Wu, Bo Zhao, Derong Liu
{"title":"LSBO-NAS: Latent Space Bayesian Optimization for Neural Architecture Search","authors":"Xuan Rao, Songyi Xiao, Jiaxin Li, Qiuye Wu, Bo Zhao, Derong Liu","doi":"10.1109/ICCR55715.2022.10053904","DOIUrl":null,"url":null,"abstract":"From the perspective of data stream, neural architecture search (NAS) can be formulated as a graph optimization problem. However, many state-of-the-art black-box optimization algorithms, such as Bayesian optimization and simulated annealing, operate in continuous space primarily, which does not match the NAS optimization due to the discreteness of graph structures. To tackle this problem, the latent space Bayesian optimization NAS (LSBO-NAS) algorithm is developed in this paper. In LSBO-NAS, the neural architectures are represented as sequences, and a variational auto-encoder (VAE) is trained to convert the discrete search space of NAS into a continuous latent space by learning the continuous representation of neural architectures. Hereafter, a Bayesian optimization (BO) algorithm, i.e., the tree-structure parzen estimator (TPE) algorithm, is developed to obtain admirable neural architectures. The optimization loop of LSBO-NAS consists of two stages. In the first stage, the BO algorithm generates a preferable architecture representation according to its search strategy. In the second stage, the decoder of VAE decodes the representation into a discrete neural architecture, whose performance evaluation is regarded as the feedback signal for the BO algorithm. The effectiveness of the developed LSBO-NAS is demonstrated on the NAS-Bench-301 benchmark, where the LSBO-NAS achieves a better performance than several NAS baselines.","PeriodicalId":441511,"journal":{"name":"2022 4th International Conference on Control and Robotics (ICCR)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 4th International Conference on Control and Robotics (ICCR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCR55715.2022.10053904","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
From the perspective of data stream, neural architecture search (NAS) can be formulated as a graph optimization problem. However, many state-of-the-art black-box optimization algorithms, such as Bayesian optimization and simulated annealing, operate in continuous space primarily, which does not match the NAS optimization due to the discreteness of graph structures. To tackle this problem, the latent space Bayesian optimization NAS (LSBO-NAS) algorithm is developed in this paper. In LSBO-NAS, the neural architectures are represented as sequences, and a variational auto-encoder (VAE) is trained to convert the discrete search space of NAS into a continuous latent space by learning the continuous representation of neural architectures. Hereafter, a Bayesian optimization (BO) algorithm, i.e., the tree-structure parzen estimator (TPE) algorithm, is developed to obtain admirable neural architectures. The optimization loop of LSBO-NAS consists of two stages. In the first stage, the BO algorithm generates a preferable architecture representation according to its search strategy. In the second stage, the decoder of VAE decodes the representation into a discrete neural architecture, whose performance evaluation is regarded as the feedback signal for the BO algorithm. The effectiveness of the developed LSBO-NAS is demonstrated on the NAS-Bench-301 benchmark, where the LSBO-NAS achieves a better performance than several NAS baselines.