无需直线搜索的非精确正则近似牛顿法

IF 1.6 2区 数学 Q2 MATHEMATICS, APPLIED Computational Optimization and Applications Pub Date : 2024-08-16 DOI:10.1007/s10589-024-00600-9
Simeon vom Dahl, Christian Kanzow
{"title":"无需直线搜索的非精确正则近似牛顿法","authors":"Simeon vom Dahl, Christian Kanzow","doi":"10.1007/s10589-024-00600-9","DOIUrl":null,"url":null,"abstract":"<p>In this paper, we introduce an inexact regularized proximal Newton method (IRPNM) that does not require any line search. The method is designed to minimize the sum of a twice continuously differentiable function <i>f</i> and a convex (possibly non-smooth and extended-valued) function <span>\\(\\varphi \\)</span>. Instead of controlling a step size by a line search procedure, we update the regularization parameter in a suitable way, based on the success of the previous iteration. The global convergence of the sequence of iterations and its superlinear convergence rate under a local Hölderian error bound assumption are shown. Notably, these convergence results are obtained without requiring a global Lipschitz property for <span>\\( \\nabla f \\)</span>, which, to the best of the authors’ knowledge, is a novel contribution for proximal Newton methods. To highlight the efficiency of our approach, we provide numerical comparisons with an IRPNM using a line search globalization and a modern FISTA-type method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"29 1","pages":""},"PeriodicalIF":1.6000,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An inexact regularized proximal Newton method without line search\",\"authors\":\"Simeon vom Dahl, Christian Kanzow\",\"doi\":\"10.1007/s10589-024-00600-9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>In this paper, we introduce an inexact regularized proximal Newton method (IRPNM) that does not require any line search. The method is designed to minimize the sum of a twice continuously differentiable function <i>f</i> and a convex (possibly non-smooth and extended-valued) function <span>\\\\(\\\\varphi \\\\)</span>. Instead of controlling a step size by a line search procedure, we update the regularization parameter in a suitable way, based on the success of the previous iteration. The global convergence of the sequence of iterations and its superlinear convergence rate under a local Hölderian error bound assumption are shown. Notably, these convergence results are obtained without requiring a global Lipschitz property for <span>\\\\( \\\\nabla f \\\\)</span>, which, to the best of the authors’ knowledge, is a novel contribution for proximal Newton methods. To highlight the efficiency of our approach, we provide numerical comparisons with an IRPNM using a line search globalization and a modern FISTA-type method.</p>\",\"PeriodicalId\":55227,\"journal\":{\"name\":\"Computational Optimization and Applications\",\"volume\":\"29 1\",\"pages\":\"\"},\"PeriodicalIF\":1.6000,\"publicationDate\":\"2024-08-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computational Optimization and Applications\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1007/s10589-024-00600-9\",\"RegionNum\":2,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"MATHEMATICS, APPLIED\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computational Optimization and Applications","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s10589-024-00600-9","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0

摘要

在本文中,我们介绍了一种不需要任何直线搜索的非精确正则化近似牛顿法(IRPNM)。该方法旨在最小化两次连续可微分函数 f 与凸(可能是非光滑和扩展值)函数 \(\varphi \)之和。我们不是通过直线搜索程序来控制步长,而是根据前一次迭代的成功率,以适当的方式更新正则化参数。在局部霍尔德误差约束假设下,迭代序列的全局收敛性及其超线性收敛率得到了证明。值得注意的是,这些收敛结果是在不要求 \( \nabla f \) 的全局 Lipschitz 属性的情况下获得的,据作者所知,这是近似牛顿方法的一个新贡献。为了突出我们方法的效率,我们提供了使用线搜索全局化的 IRPNM 和现代 FISTA 类型方法的数值比较。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

摘要图片

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
An inexact regularized proximal Newton method without line search

In this paper, we introduce an inexact regularized proximal Newton method (IRPNM) that does not require any line search. The method is designed to minimize the sum of a twice continuously differentiable function f and a convex (possibly non-smooth and extended-valued) function \(\varphi \). Instead of controlling a step size by a line search procedure, we update the regularization parameter in a suitable way, based on the success of the previous iteration. The global convergence of the sequence of iterations and its superlinear convergence rate under a local Hölderian error bound assumption are shown. Notably, these convergence results are obtained without requiring a global Lipschitz property for \( \nabla f \), which, to the best of the authors’ knowledge, is a novel contribution for proximal Newton methods. To highlight the efficiency of our approach, we provide numerical comparisons with an IRPNM using a line search globalization and a modern FISTA-type method.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
3.70
自引率
9.10%
发文量
91
审稿时长
10 months
期刊介绍: Computational Optimization and Applications is a peer reviewed journal that is committed to timely publication of research and tutorial papers on the analysis and development of computational algorithms and modeling technology for optimization. Algorithms either for general classes of optimization problems or for more specific applied problems are of interest. Stochastic algorithms as well as deterministic algorithms will be considered. Papers that can provide both theoretical analysis, along with carefully designed computational experiments, are particularly welcome. Topics of interest include, but are not limited to the following: Large Scale Optimization, Unconstrained Optimization, Linear Programming, Quadratic Programming Complementarity Problems, and Variational Inequalities, Constrained Optimization, Nondifferentiable Optimization, Integer Programming, Combinatorial Optimization, Stochastic Optimization, Multiobjective Optimization, Network Optimization, Complexity Theory, Approximations and Error Analysis, Parametric Programming and Sensitivity Analysis, Parallel Computing, Distributed Computing, and Vector Processing, Software, Benchmarks, Numerical Experimentation and Comparisons, Modelling Languages and Systems for Optimization, Automatic Differentiation, Applications in Engineering, Finance, Optimal Control, Optimal Design, Operations Research, Transportation, Economics, Communications, Manufacturing, and Management Science.
期刊最新文献
A family of conjugate gradient methods with guaranteed positiveness and descent for vector optimization Convergence of a quasi-Newton method for solving systems of nonlinear underdetermined equations Scaled-PAKKT sequential optimality condition for multiobjective problems and its application to an Augmented Lagrangian method A Newton-CG based barrier-augmented Lagrangian method for general nonconvex conic optimization Robust approximation of chance constrained optimization with polynomial perturbation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1