Modification of the Marquardt method for training a neural network predictor in eddy viscosity models

V. V. Pekunov
{"title":"Modification of the Marquardt method for training a neural network predictor in eddy viscosity models","authors":"V. V. Pekunov","doi":"10.25136/2644-5522.2021.1.36059","DOIUrl":null,"url":null,"abstract":"\n The subject of this article is the numerical optimization techniques used in training neural networks that serve as predicate components in certain modern eddy viscosity models. Qualitative solution to the problem of training (minimization of the functional of neural network offsets) often requires significant computational costs, which necessitates to increase the speed of such training based on combination of numerical methods and parallelization of calculations. The Marquardt method draws particular interest, as it contains  the parameter that allows speeding up the solution by switching the method from the descent away from the solution to the Newton’s method of approximate solution. The article offers modification of the Marquardt method, which uses the limited series of random samples for improving the current point and calculate the parameter of the method. The author demonstrate descent characteristics of the method in numerical experiments, both on the test functions of Himmelblau and Rosenbrock, as well as the actual task of training the neural network predictor applies in modeling of the turbulent flows. The use of this method may significantly speed up the training of neural network predictor in corrective models of eddy viscosity. The method is less time-consuming in comparison with random search, namely in terms of a small amount of compute kernels; however, it provides solution that is close to the result of random search and is better than the original Marquardt method.\n","PeriodicalId":351916,"journal":{"name":"Кибернетика и программирование","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Кибернетика и программирование","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.25136/2644-5522.2021.1.36059","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The subject of this article is the numerical optimization techniques used in training neural networks that serve as predicate components in certain modern eddy viscosity models. Qualitative solution to the problem of training (minimization of the functional of neural network offsets) often requires significant computational costs, which necessitates to increase the speed of such training based on combination of numerical methods and parallelization of calculations. The Marquardt method draws particular interest, as it contains  the parameter that allows speeding up the solution by switching the method from the descent away from the solution to the Newton’s method of approximate solution. The article offers modification of the Marquardt method, which uses the limited series of random samples for improving the current point and calculate the parameter of the method. The author demonstrate descent characteristics of the method in numerical experiments, both on the test functions of Himmelblau and Rosenbrock, as well as the actual task of training the neural network predictor applies in modeling of the turbulent flows. The use of this method may significantly speed up the training of neural network predictor in corrective models of eddy viscosity. The method is less time-consuming in comparison with random search, namely in terms of a small amount of compute kernels; however, it provides solution that is close to the result of random search and is better than the original Marquardt method.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
涡流粘度模型中训练神经网络预测器的Marquardt方法的改进
本文的主题是用于训练神经网络的数值优化技术,这些神经网络在某些现代涡动黏度模型中充当谓词组件。定性解决训练问题(神经网络偏移量函数的最小化)往往需要大量的计算成本,这就需要在数值方法和计算并行化相结合的基础上提高训练速度。马夸特方法引起了特别的兴趣,因为它包含了一个参数,可以通过将方法从解的下降转换为近似解的牛顿方法来加速解的速度。本文提出了对Marquardt方法的改进,利用有限的随机样本序列对当前点进行改进,并计算方法的参数。在Himmelblau和Rosenbrock测试函数的数值实验中,以及训练神经网络预测器应用于湍流建模的实际任务中,证明了该方法的下降特性。该方法的应用可以显著加快涡流粘度校正模型中神经网络预测器的训练速度。与随机搜索相比,该方法耗时更少,即计算核数量较少;然而,它提供了接近随机搜索结果的解,并且优于原来的Marquardt方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Entropy estimation of the fragments of chest X-ray images Modification of the Marquardt method for training a neural network predictor in eddy viscosity models Improved CPU load balancing for numerical solution of the tasks of continuous medium mechanics complicated by chemical kinetics Walking robots for rescue operations: overview and analysis of the existing models Design of the database structure for software optimization of operation of the stochastic multiphase systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1