神经 PDE 求解器的主动学习

Daniel Musekamp, Marimuthu Kalimuthu, David Holzmüller, Makoto Takamoto, Mathias Niepert
{"title":"神经 PDE 求解器的主动学习","authors":"Daniel Musekamp, Marimuthu Kalimuthu, David Holzmüller, Makoto Takamoto, Mathias Niepert","doi":"arxiv-2408.01536","DOIUrl":null,"url":null,"abstract":"Solving partial differential equations (PDEs) is a fundamental problem in\nengineering and science. While neural PDE solvers can be more efficient than\nestablished numerical solvers, they often require large amounts of training\ndata that is costly to obtain. Active Learning (AL) could help surrogate models\nreach the same accuracy with smaller training sets by querying classical\nsolvers with more informative initial conditions and PDE parameters. While AL\nis more common in other domains, it has yet to be studied extensively for\nneural PDE solvers. To bridge this gap, we introduce AL4PDE, a modular and\nextensible active learning benchmark. It provides multiple parametric PDEs and\nstate-of-the-art surrogate models for the solver-in-the-loop setting, enabling\nthe evaluation of existing and the development of new AL methods for PDE\nsolving. We use the benchmark to evaluate batch active learning algorithms such\nas uncertainty- and feature-based methods. We show that AL reduces the average\nerror by up to 71% compared to random sampling and significantly reduces\nworst-case errors. Moreover, AL generates similar datasets across repeated\nruns, with consistent distributions over the PDE parameters and initial\nconditions. The acquired datasets are reusable, providing benefits for\nsurrogate models not involved in the data generation.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"8 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Active Learning for Neural PDE Solvers\",\"authors\":\"Daniel Musekamp, Marimuthu Kalimuthu, David Holzmüller, Makoto Takamoto, Mathias Niepert\",\"doi\":\"arxiv-2408.01536\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Solving partial differential equations (PDEs) is a fundamental problem in\\nengineering and science. While neural PDE solvers can be more efficient than\\nestablished numerical solvers, they often require large amounts of training\\ndata that is costly to obtain. Active Learning (AL) could help surrogate models\\nreach the same accuracy with smaller training sets by querying classical\\nsolvers with more informative initial conditions and PDE parameters. While AL\\nis more common in other domains, it has yet to be studied extensively for\\nneural PDE solvers. To bridge this gap, we introduce AL4PDE, a modular and\\nextensible active learning benchmark. It provides multiple parametric PDEs and\\nstate-of-the-art surrogate models for the solver-in-the-loop setting, enabling\\nthe evaluation of existing and the development of new AL methods for PDE\\nsolving. We use the benchmark to evaluate batch active learning algorithms such\\nas uncertainty- and feature-based methods. We show that AL reduces the average\\nerror by up to 71% compared to random sampling and significantly reduces\\nworst-case errors. Moreover, AL generates similar datasets across repeated\\nruns, with consistent distributions over the PDE parameters and initial\\nconditions. The acquired datasets are reusable, providing benefits for\\nsurrogate models not involved in the data generation.\",\"PeriodicalId\":501347,\"journal\":{\"name\":\"arXiv - CS - Neural and Evolutionary Computing\",\"volume\":\"8 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Neural and Evolutionary Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.01536\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Neural and Evolutionary Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.01536","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

求解偏微分方程(PDE)是工程和科学领域的一个基本问题。虽然神经偏微分方程求解器比现有的数值求解器更高效,但它们通常需要大量的训练数据,而获取这些数据的成本很高。主动学习(Active Learning,AL)可以帮助代用模型以更小的训练集达到相同的精度,方法是用更多信息的初始条件和 PDE 参数查询经典求解器。虽然主动学习在其他领域更为常见,但在神经 PDE 求解器方面还没有广泛的研究。为了弥补这一差距,我们引入了 AL4PDE,这是一种模块化、可扩展的主动学习基准。它为解算器在环设置提供了多个参数化 PDE 和最先进的代理模型,使我们能够评估现有的 PDE 求解方法并开发新的 AL 方法。我们使用该基准来评估批量主动学习算法,如基于不确定性和特征的方法。我们的研究表明,与随机抽样相比,AL 将平均误差降低了 71%,并显著降低了最坏情况下的误差。此外,AL 还能在重复运行中生成相似的数据集,并且在 PDE 参数和初始条件上具有一致的分布。获得的数据集可重复使用,为未参与数据生成的代用模型带来了好处。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Active Learning for Neural PDE Solvers
Solving partial differential equations (PDEs) is a fundamental problem in engineering and science. While neural PDE solvers can be more efficient than established numerical solvers, they often require large amounts of training data that is costly to obtain. Active Learning (AL) could help surrogate models reach the same accuracy with smaller training sets by querying classical solvers with more informative initial conditions and PDE parameters. While AL is more common in other domains, it has yet to be studied extensively for neural PDE solvers. To bridge this gap, we introduce AL4PDE, a modular and extensible active learning benchmark. It provides multiple parametric PDEs and state-of-the-art surrogate models for the solver-in-the-loop setting, enabling the evaluation of existing and the development of new AL methods for PDE solving. We use the benchmark to evaluate batch active learning algorithms such as uncertainty- and feature-based methods. We show that AL reduces the average error by up to 71% compared to random sampling and significantly reduces worst-case errors. Moreover, AL generates similar datasets across repeated runs, with consistent distributions over the PDE parameters and initial conditions. The acquired datasets are reusable, providing benefits for surrogate models not involved in the data generation.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Hardware-Friendly Implementation of Physical Reservoir Computing with CMOS-based Time-domain Analog Spiking Neurons Self-Contrastive Forward-Forward Algorithm Bio-Inspired Mamba: Temporal Locality and Bioplausible Learning in Selective State Space Models PReLU: Yet Another Single-Layer Solution to the XOR Problem Inferno: An Extensible Framework for Spiking Neural Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1