Manifold learning for accelerating coarse-grained optimization

IF 16.4 1区 化学 Q1 CHEMISTRY, MULTIDISCIPLINARY Accounts of Chemical Research Pub Date : 2020-01-10 DOI:10.3934/jcd.2020021
D. Pozharskiy, Noah J. Wichrowski, A. Duncan, G. Pavliotis, I. Kevrekidis
{"title":"Manifold learning for accelerating coarse-grained optimization","authors":"D. Pozharskiy, Noah J. Wichrowski, A. Duncan, G. Pavliotis, I. Kevrekidis","doi":"10.3934/jcd.2020021","DOIUrl":null,"url":null,"abstract":"Algorithms proposed for solving high-dimensional optimization problems with no derivative information frequently encounter the \"curse of dimensionality,\" becoming ineffective as the dimension of the parameter space grows. One feature of a subclass of such problems that are effectively low-dimensional is that only a few parameters (or combinations thereof) are important for the optimization and must be explored in detail. Knowing these parameters/ combinations in advance would greatly simplify the problem and its solution. We propose the data-driven construction of an effective (coarse-grained, \"trend\") optimizer, based on data obtained from ensembles of brief simulation bursts with an \"inner\" optimization algorithm, that has the potential to accelerate the exploration of the parameter space. The trajectories of this \"effective optimizer\" quickly become attracted onto a slow manifold parameterized by the few relevant parameter combinations. We obtain the parameterization of this low-dimensional, effective optimization manifold on the fly using data mining/manifold learning techniques on the results of simulation (inner optimizer iteration) burst ensembles and exploit it locally to \"jump\" forward along this manifold. As a result, we can bias the exploration of the parameter space towards the few, important directions and, through this \"wrapper algorithm,\" speed up the convergence of traditional optimization algorithms.","PeriodicalId":1,"journal":{"name":"Accounts of Chemical Research","volume":null,"pages":null},"PeriodicalIF":16.4000,"publicationDate":"2020-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Accounts of Chemical Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3934/jcd.2020021","RegionNum":1,"RegionCategory":"化学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CHEMISTRY, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 6

Abstract

Algorithms proposed for solving high-dimensional optimization problems with no derivative information frequently encounter the "curse of dimensionality," becoming ineffective as the dimension of the parameter space grows. One feature of a subclass of such problems that are effectively low-dimensional is that only a few parameters (or combinations thereof) are important for the optimization and must be explored in detail. Knowing these parameters/ combinations in advance would greatly simplify the problem and its solution. We propose the data-driven construction of an effective (coarse-grained, "trend") optimizer, based on data obtained from ensembles of brief simulation bursts with an "inner" optimization algorithm, that has the potential to accelerate the exploration of the parameter space. The trajectories of this "effective optimizer" quickly become attracted onto a slow manifold parameterized by the few relevant parameter combinations. We obtain the parameterization of this low-dimensional, effective optimization manifold on the fly using data mining/manifold learning techniques on the results of simulation (inner optimizer iteration) burst ensembles and exploit it locally to "jump" forward along this manifold. As a result, we can bias the exploration of the parameter space towards the few, important directions and, through this "wrapper algorithm," speed up the convergence of traditional optimization algorithms.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
求解无导数信息的高维优化问题的算法经常遇到“维数诅咒”,随着参数空间维数的增加而变得无效。有效低维问题的子类的一个特征是,只有几个参数(或它们的组合)对优化是重要的,必须详细探索。事先知道这些参数/组合将大大简化问题及其解决方案。我们提出了一个有效的(粗粒度的,“趋势”)优化器的数据驱动结构,基于从具有“内部”优化算法的简短模拟爆发集合中获得的数据,这有可能加速对参数空间的探索。这个“有效优化器”的轨迹很快被吸引到一个由少数相关参数组合参数化的慢流形上。我们利用数据挖掘/流形学习技术对模拟(内部优化器迭代)突发集合的结果动态地获得了这种低维有效优化流形的参数化,并利用它在局部沿着该流形向前“跳跃”。因此,我们可以将参数空间的探索偏向于少数重要的方向,并通过这种“包装算法”加快传统优化算法的收敛速度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Accounts of Chemical Research
Accounts of Chemical Research 化学-化学综合
CiteScore
31.40
自引率
1.10%
发文量
312
审稿时长
2 months
期刊介绍: Accounts of Chemical Research presents short, concise and critical articles offering easy-to-read overviews of basic research and applications in all areas of chemistry and biochemistry. These short reviews focus on research from the author’s own laboratory and are designed to teach the reader about a research project. In addition, Accounts of Chemical Research publishes commentaries that give an informed opinion on a current research problem. Special Issues online are devoted to a single topic of unusual activity and significance. Accounts of Chemical Research replaces the traditional article abstract with an article "Conspectus." These entries synopsize the research affording the reader a closer look at the content and significance of an article. Through this provision of a more detailed description of the article contents, the Conspectus enhances the article's discoverability by search engines and the exposure for the research.
期刊最新文献
The change process questionnaire (CPQ): A psychometric validation. Differential Costs of Raising Grandchildren on Older Mother-Adult Child Relations in Black and White Families. Does Resilience Mediate the Relationship Between Negative Self-Image and Psychological Distress in Middle-Aged and Older Gay and Bisexual Men? Intergenerational Relations and Well-being Among Older Middle Eastern/Arab American Immigrants During the COVID-19 Pandemic. Caregiving Appraisals and Emotional Valence: Moderating Effects of Activity Participation.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1