基于逆修正微分方程的误差分析,利用线性多步骤方法和深度学习发现动力学规律

IF 2.8 2区 数学 Q1 MATHEMATICS, APPLIED SIAM Journal on Numerical Analysis Pub Date : 2024-09-04 DOI:10.1137/22m152373x
Aiqing Zhu, Sidi Wu, Yifa Tang
{"title":"基于逆修正微分方程的误差分析,利用线性多步骤方法和深度学习发现动力学规律","authors":"Aiqing Zhu, Sidi Wu, Yifa Tang","doi":"10.1137/22m152373x","DOIUrl":null,"url":null,"abstract":"SIAM Journal on Numerical Analysis, Volume 62, Issue 5, Page 2087-2120, October 2024. <br/> Abstract. Along with the practical success of the discovery of dynamics using deep learning, the theoretical analysis of this approach has attracted increasing attention. Prior works have established the grid error estimation with auxiliary conditions for the discovery of dynamics using linear multistep methods and deep learning. And we extend the existing error analysis in this work. We first introduce the concept of inverse modified differential equations (IMDE) for linear multistep methods and show that the learned model returns a close approximation of the IMDE. Based on the IMDE, we prove that the error between the discovered system and the target system is bounded by the sum of the LMM discretization error and the learning loss. Furthermore, the learning loss is quantified by combining the approximation and generalization theories of neural networks, and thereby we obtain the priori error estimates. Several numerical experiments are performed to verify the theoretical analysis.","PeriodicalId":49527,"journal":{"name":"SIAM Journal on Numerical Analysis","volume":null,"pages":null},"PeriodicalIF":2.8000,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Error Analysis Based on Inverse Modified Differential Equations for Discovery of Dynamics Using Linear Multistep Methods and Deep Learning\",\"authors\":\"Aiqing Zhu, Sidi Wu, Yifa Tang\",\"doi\":\"10.1137/22m152373x\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"SIAM Journal on Numerical Analysis, Volume 62, Issue 5, Page 2087-2120, October 2024. <br/> Abstract. Along with the practical success of the discovery of dynamics using deep learning, the theoretical analysis of this approach has attracted increasing attention. Prior works have established the grid error estimation with auxiliary conditions for the discovery of dynamics using linear multistep methods and deep learning. And we extend the existing error analysis in this work. We first introduce the concept of inverse modified differential equations (IMDE) for linear multistep methods and show that the learned model returns a close approximation of the IMDE. Based on the IMDE, we prove that the error between the discovered system and the target system is bounded by the sum of the LMM discretization error and the learning loss. Furthermore, the learning loss is quantified by combining the approximation and generalization theories of neural networks, and thereby we obtain the priori error estimates. Several numerical experiments are performed to verify the theoretical analysis.\",\"PeriodicalId\":49527,\"journal\":{\"name\":\"SIAM Journal on Numerical Analysis\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.8000,\"publicationDate\":\"2024-09-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"SIAM Journal on Numerical Analysis\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1137/22m152373x\",\"RegionNum\":2,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS, APPLIED\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"SIAM Journal on Numerical Analysis","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1137/22m152373x","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0

摘要

SIAM 数值分析期刊》,第 62 卷第 5 期,第 2087-2120 页,2024 年 10 月。 摘要随着利用深度学习发现动力学的实践成功,这种方法的理论分析也引起了越来越多的关注。之前的工作建立了利用线性多步方法和深度学习发现动力学的网格误差估计与辅助条件。而我们在这项工作中扩展了现有的误差分析。我们首先为线性多步方法引入了逆修正微分方程(IMDE)的概念,并证明学习模型返回的是 IMDE 的近似值。基于 IMDE,我们证明了所发现的系统与目标系统之间的误差以 LMM 离散化误差和学习损失之和为界。此外,我们还结合神经网络的近似和泛化理论对学习损失进行了量化,从而获得了先验误差估计值。我们进行了一些数值实验来验证理论分析。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Error Analysis Based on Inverse Modified Differential Equations for Discovery of Dynamics Using Linear Multistep Methods and Deep Learning
SIAM Journal on Numerical Analysis, Volume 62, Issue 5, Page 2087-2120, October 2024.
Abstract. Along with the practical success of the discovery of dynamics using deep learning, the theoretical analysis of this approach has attracted increasing attention. Prior works have established the grid error estimation with auxiliary conditions for the discovery of dynamics using linear multistep methods and deep learning. And we extend the existing error analysis in this work. We first introduce the concept of inverse modified differential equations (IMDE) for linear multistep methods and show that the learned model returns a close approximation of the IMDE. Based on the IMDE, we prove that the error between the discovered system and the target system is bounded by the sum of the LMM discretization error and the learning loss. Furthermore, the learning loss is quantified by combining the approximation and generalization theories of neural networks, and thereby we obtain the priori error estimates. Several numerical experiments are performed to verify the theoretical analysis.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
4.80
自引率
6.90%
发文量
110
审稿时长
4-8 weeks
期刊介绍: SIAM Journal on Numerical Analysis (SINUM) contains research articles on the development and analysis of numerical methods. Topics include the rigorous study of convergence of algorithms, their accuracy, their stability, and their computational complexity. Also included are results in mathematical analysis that contribute to algorithm analysis, and computational results that demonstrate algorithm behavior and applicability.
期刊最新文献
How Sharp Are Error Bounds? –Lower Bounds on Quadrature Worst-Case Errors for Analytic Functions– Fractal Multiquadric Interpolation Functions High Order Biorthogonal Functions in [math](curl) Convergence Analysis of the Parareal Algorithm with Nonuniform Fine Time Grid Numerical Reconstruction of Diffusion and Potential Coefficients from Two Observations: Decoupled Recovery and Error Estimates
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1