首页 > 最新文献

Computational Optimization and Applications最新文献

英文 中文
A family of conjugate gradient methods with guaranteed positiveness and descent for vector optimization 保证正向性和下降的共轭梯度法系列,用于矢量优化
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-09-17 DOI: 10.1007/s10589-024-00609-0
Qing-Rui He, Sheng-Jie Li, Bo-Ya Zhang, Chun-Rong Chen

In this paper, we seek a new modification way to ensure the positiveness of the conjugate parameter and, based on the Dai-Yuan (DY) method in the vector setting, propose an associated family of conjugate gradient (CG) methods with guaranteed descent for solving unconstrained vector optimization problems. Several special members of the family are analyzed and the (sufficient) descent condition is established for them (in the vector sense). Under mild conditions, a general convergence result for the CG methods with specific parameters is presented, which, in particular, covers the global convergence of the aforementioned members. Furthermore, for the purpose of comparison, we then consider the direct extension versions of some Dai-Yuan type methods which are obtained by modifying the DY method of the scalar case. These vector extensions can retrieve the classical parameters in the scalar minimization case and their descent property and global convergence are also studied under mild assumptions. Finally, numerical experiments are given to illustrate the practical behavior of all proposed methods.

在本文中,我们寻求了一种新的修正方法来确保共轭参数的正向性,并基于矢量环境下的戴元(DY)方法,提出了一个相关的共轭梯度(CG)方法族,该方法具有保证下降的特性,可用于求解无约束矢量优化问题。分析了该族的几个特殊成员,并为它们建立了(向量意义上的)(充分)下降条件。在温和条件下,提出了具有特定参数的 CG 方法的一般收敛结果,特别是涵盖了上述成员的全局收敛性。此外,为了进行比较,我们还考虑了一些戴元类方法的直接扩展版本,它们是通过修改标量情况下的 DY 方法而获得的。这些矢量扩展方法可以检索标量最小化情况下的经典参数,并在温和的假设条件下研究了它们的下降特性和全局收敛性。最后,还给出了数值实验来说明所有建议方法的实用性。
{"title":"A family of conjugate gradient methods with guaranteed positiveness and descent for vector optimization","authors":"Qing-Rui He, Sheng-Jie Li, Bo-Ya Zhang, Chun-Rong Chen","doi":"10.1007/s10589-024-00609-0","DOIUrl":"https://doi.org/10.1007/s10589-024-00609-0","url":null,"abstract":"<p>In this paper, we seek a new modification way to ensure the positiveness of the conjugate parameter and, based on the Dai-Yuan (DY) method in the vector setting, propose an associated family of conjugate gradient (CG) methods with guaranteed descent for solving unconstrained vector optimization problems. Several special members of the family are analyzed and the (sufficient) descent condition is established for them (in the vector sense). Under mild conditions, a general convergence result for the CG methods with specific parameters is presented, which, in particular, covers the global convergence of the aforementioned members. Furthermore, for the purpose of comparison, we then consider the direct extension versions of some Dai-Yuan type methods which are obtained by modifying the DY method of the scalar case. These vector extensions can retrieve the classical parameters in the scalar minimization case and their descent property and global convergence are also studied under mild assumptions. Finally, numerical experiments are given to illustrate the practical behavior of all proposed methods.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"47 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142256765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convergence of a quasi-Newton method for solving systems of nonlinear underdetermined equations 求解非线性欠定方程组的准牛顿法的收敛性
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-09-06 DOI: 10.1007/s10589-024-00606-3
N. Vater, A. Borzì

The development and convergence analysis of a quasi-Newton method for the solution of systems of nonlinear underdetermined equations is investigated. These equations arise in many application fields, e.g., supervised learning of large overparameterised neural networks, which require the development of efficient methods with guaranteed convergence. In this paper, a new approach for the computation of the Moore–Penrose inverse of the approximate Jacobian coming from the Broyden update is presented and a semi-local convergence result for a damped quasi-Newton method is proved. The theoretical results are illustrated in detail for the case of systems of multidimensional quadratic equations, and validated in the context of eigenvalue problems and supervised learning of overparameterised neural networks.

研究了求解非线性欠定方程系统的准牛顿方法的开发和收敛分析。这些方程出现在许多应用领域,例如大型过参数化神经网络的监督学习,这就需要开发具有收敛性保证的高效方法。本文提出了一种计算来自布洛伊登更新的近似雅各布逆的摩尔-彭罗斯逆的新方法,并证明了阻尼准牛顿方法的半局部收敛结果。理论结果详细说明了多维二次方程组的情况,并在特征值问题和过参数化神经网络的监督学习中得到了验证。
{"title":"Convergence of a quasi-Newton method for solving systems of nonlinear underdetermined equations","authors":"N. Vater, A. Borzì","doi":"10.1007/s10589-024-00606-3","DOIUrl":"https://doi.org/10.1007/s10589-024-00606-3","url":null,"abstract":"<p>The development and convergence analysis of a quasi-Newton method for the solution of systems of nonlinear underdetermined equations is investigated. These equations arise in many application fields, e.g., supervised learning of large overparameterised neural networks, which require the development of efficient methods with guaranteed convergence. In this paper, a new approach for the computation of the Moore–Penrose inverse of the approximate Jacobian coming from the Broyden update is presented and a semi-local convergence result for a damped quasi-Newton method is proved. The theoretical results are illustrated in detail for the case of systems of multidimensional quadratic equations, and validated in the context of eigenvalue problems and supervised learning of overparameterised neural networks.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"91 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142199602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scaled-PAKKT sequential optimality condition for multiobjective problems and its application to an Augmented Lagrangian method 多目标问题的 Scaled-PAKKT 顺序最优条件及其在增强拉格朗日法中的应用
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-09-02 DOI: 10.1007/s10589-024-00605-4
G. A. Carrizo, N. S. Fazzio, M. D. Sánchez, M. L. Schuverdt

Based on the recently introduced Scaled Positive Approximate Karush–Kuhn–Tucker condition for single objective problems, we derive a sequential necessary optimality condition for multiobjective problems with equality and inequality constraints as well as additional abstract set constraints. These necessary sequential optimality conditions for multiobjective problems are subject to the same requirements as ordinary (pointwise) optimization conditions: we show that the updated Scaled Positive Approximate Karush–Kuhn–Tucker condition is necessary for a local weak Pareto point to the problem. Furthermore, we propose a variant of the classical Augmented Lagrangian method for multiobjective problems. Our theoretical framework does not require any scalarization. We also discuss the convergence properties of our algorithm with regard to feasibility and global optimality without any convexity assumption. Finally, some numerical results are given to illustrate the practical viability of the method.

基于最近推出的针对单目标问题的按比例正向近似卡罗什-库恩-塔克条件,我们推导出了针对具有相等和不相等约束以及额外抽象集约束的多目标问题的顺序必要最优条件。多目标问题的这些必要顺序优化条件与普通(点式)优化条件的要求相同:我们证明,更新后的正向近似卡罗什-库恩-塔克条件对于问题的局部弱帕累托点是必要的。此外,我们还为多目标问题提出了经典的增量拉格朗日法的变体。我们的理论框架不需要任何标量化。我们还讨论了我们的算法在可行性和全局最优性方面的收敛特性,而无需任何凸性假设。最后,我们给出了一些数值结果,以说明该方法的实际可行性。
{"title":"Scaled-PAKKT sequential optimality condition for multiobjective problems and its application to an Augmented Lagrangian method","authors":"G. A. Carrizo, N. S. Fazzio, M. D. Sánchez, M. L. Schuverdt","doi":"10.1007/s10589-024-00605-4","DOIUrl":"https://doi.org/10.1007/s10589-024-00605-4","url":null,"abstract":"<p>Based on the recently introduced Scaled Positive Approximate Karush–Kuhn–Tucker condition for single objective problems, we derive a sequential necessary optimality condition for multiobjective problems with equality and inequality constraints as well as additional abstract set constraints. These necessary sequential optimality conditions for multiobjective problems are subject to the same requirements as ordinary (pointwise) optimization conditions: we show that the updated Scaled Positive Approximate Karush–Kuhn–Tucker condition is necessary for a local weak Pareto point to the problem. Furthermore, we propose a variant of the classical Augmented Lagrangian method for multiobjective problems. Our theoretical framework does not require any scalarization. We also discuss the convergence properties of our algorithm with regard to feasibility and global optimality without any convexity assumption. Finally, some numerical results are given to illustrate the practical viability of the method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"59 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142199603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Newton-CG based barrier-augmented Lagrangian method for general nonconvex conic optimization 基于牛顿-CG 的一般非凸圆锥优化的障碍增强拉格朗日方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-08-30 DOI: 10.1007/s10589-024-00603-6
Chuan He, Heng Huang, Zhaosong Lu

In this paper we consider finding an approximate second-order stationary point (SOSP) of general nonconvex conic optimization that minimizes a twice differentiable function subject to nonlinear equality constraints and also a convex conic constraint. In particular, we propose a Newton-conjugate gradient (Newton-CG) based barrier-augmented Lagrangian method for finding an approximate SOSP of this problem. Under some mild assumptions, we show that our method enjoys a total inner iteration complexity of ({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-11/2})) and an operation complexity of ({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-11/2}min {n,epsilon ^{-5/4}})) for finding an ((epsilon ,sqrt{epsilon }))-SOSP of general nonconvex conic optimization with high probability. Moreover, under a constraint qualification, these complexity bounds are improved to ({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-7/2})) and ({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-7/2}min {n,epsilon ^{-3/4}})), respectively. To the best of our knowledge, this is the first study on the complexity of finding an approximate SOSP of general nonconvex conic optimization. Preliminary numerical results are presented to demonstrate superiority of the proposed method over first-order methods in terms of solution quality.

在本文中,我们考虑寻找一般非凸圆锥优化问题的近似二阶静止点(SOSP),即在非线性相等约束和凸圆锥约束下最小化二次微分函数。我们特别提出了一种基于牛顿-共轭梯度(Newton-CG)的障碍增量拉格朗日方法,用于寻找该问题的近似 SOSP。在一些温和的假设条件下,我们证明了我们的方法内部迭代总复杂度为({widetilde{{mathrm{mathcal {O}},}}}}(epsilon ^{-11/2})),运算复杂度为({widetilde{{mathrm{mathcal {O}},}}}}(epsilon ^{-11/2}))、((epsilon,sqrtepsilon }))-SOSP的一般非凸圆锥优化的高概率。此外,在一个约束条件下,这些复杂度边界被改进为({widetilde{{,mathrmmathcal {O}}、}}}}(epsilon ^{-7/2})) 和 ({widetilde{{,mathrm{mathcal {O}},}}}}(epsilon ^{-7/2}min {n,epsilon^{-3/4}}))。据我们所知,这是首次研究一般非凸圆锥优化的近似 SOSP 的复杂性。本文给出了初步的数值结果,证明了所提出的方法在求解质量上优于一阶方法。
{"title":"A Newton-CG based barrier-augmented Lagrangian method for general nonconvex conic optimization","authors":"Chuan He, Heng Huang, Zhaosong Lu","doi":"10.1007/s10589-024-00603-6","DOIUrl":"https://doi.org/10.1007/s10589-024-00603-6","url":null,"abstract":"<p>In this paper we consider finding an approximate second-order stationary point (SOSP) of general nonconvex conic optimization that minimizes a twice differentiable function subject to nonlinear equality constraints and also a convex conic constraint. In particular, we propose a Newton-conjugate gradient (Newton-CG) based barrier-augmented Lagrangian method for finding an approximate SOSP of this problem. Under some mild assumptions, we show that our method enjoys a total inner iteration complexity of <span>({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-11/2}))</span> and an operation complexity of <span>({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-11/2}min {n,epsilon ^{-5/4}}))</span> for finding an <span>((epsilon ,sqrt{epsilon }))</span>-SOSP of general nonconvex conic optimization with high probability. Moreover, under a constraint qualification, these complexity bounds are improved to <span>({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-7/2}))</span> and <span>({widetilde{{{,mathrm{mathcal {O}},}}}}(epsilon ^{-7/2}min {n,epsilon ^{-3/4}}))</span>, respectively. To the best of our knowledge, this is the first study on the complexity of finding an approximate SOSP of general nonconvex conic optimization. Preliminary numerical results are presented to demonstrate superiority of the proposed method over first-order methods in terms of solution quality.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"165 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142199604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust approximation of chance constrained optimization with polynomial perturbation 用多项式扰动对机会约束优化进行稳健逼近
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-08-28 DOI: 10.1007/s10589-024-00602-7
Bo Rao, Liu Yang, Suhan Zhong, Guangming Zhou

This paper proposes a robust approximation method for solving chance constrained optimization (CCO) of polynomials. Assume the CCO is defined with an individual chance constraint that is affine in the decision variables. We construct a robust approximation by replacing the chance constraint with a robust constraint over an uncertainty set. When the objective function is linear or SOS-convex, the robust approximation can be equivalently transformed into linear conic optimization. Semidefinite relaxation algorithms are proposed to solve these linear conic transformations globally and their convergent properties are studied. We also introduce a heuristic method to find efficient uncertainty sets such that optimizers of the robust approximation are feasible to the original problem. Numerical experiments are given to show the efficiency of our method.

本文提出了一种求解多项式偶然约束优化(CCO)的稳健近似方法。假设 CCO 的定义包含一个单独的偶然约束,该偶然约束在决策变量中是仿射的。我们用不确定性集上的稳健约束来替代偶然约束,从而构建稳健近似法。当目标函数为线性或 SOS-凸时,稳健近似可等价转化为线性圆锥优化。我们提出了全局求解这些线性圆锥变换的半有限松弛算法,并对其收敛特性进行了研究。我们还引入了一种启发式方法来寻找有效的不确定性集,从而使鲁棒性近似的优化器对原始问题是可行的。我们给出了数值实验来证明我们方法的效率。
{"title":"Robust approximation of chance constrained optimization with polynomial perturbation","authors":"Bo Rao, Liu Yang, Suhan Zhong, Guangming Zhou","doi":"10.1007/s10589-024-00602-7","DOIUrl":"https://doi.org/10.1007/s10589-024-00602-7","url":null,"abstract":"<p>This paper proposes a robust approximation method for solving chance constrained optimization (CCO) of polynomials. Assume the CCO is defined with an individual chance constraint that is affine in the decision variables. We construct a robust approximation by replacing the chance constraint with a robust constraint over an uncertainty set. When the objective function is linear or SOS-convex, the robust approximation can be equivalently transformed into linear conic optimization. Semidefinite relaxation algorithms are proposed to solve these linear conic transformations globally and their convergent properties are studied. We also introduce a heuristic method to find efficient uncertainty sets such that optimizers of the robust approximation are feasible to the original problem. Numerical experiments are given to show the efficiency of our method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"42 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142199608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A power-like method for finding the spectral radius of a weakly irreducible nonnegative symmetric tensor 求弱不可还原非负对称张量谱半径的类幂方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-08-17 DOI: 10.1007/s10589-024-00601-8
Xueli Bai, Dong-Hui Li, Lei Wu, Jiefeng Xu

The Perron–Frobenius theorem says that the spectral radius of a weakly irreducible nonnegative tensor is the unique positive eigenvalue corresponding to a positive eigenvector. With this fact in mind, the purpose of this paper is to find the spectral radius and its corresponding positive eigenvector of a weakly irreducible nonnegative symmetric tensor. By transforming the eigenvalue problem into an equivalent problem of minimizing a concave function on a closed convex set, we derive a simpler and cheaper iterative method called power-like method, which is well-defined. Furthermore, we show that both sequences of the eigenvalue estimates and the eigenvector evaluations generated by the power-like method Q-linearly converge to the spectral radius and its corresponding eigenvector, respectively. To accelerate the method, we introduce a line search technique. The improved method retains the same convergence property as the original version. Plentiful numerical results show that the improved method performs quite well.

Perron-Frobenius 定理指出,弱不可还原非负张量的谱半径是与正特征向量相对应的唯一正特征值。考虑到这一事实,本文的目的是找出弱不可还原非负对称张量的谱半径及其对应的正特征向量。通过将特征值问题转化为在封闭凸集上最小化凹函数的等价问题,我们推导出了一种更简单、更便宜的迭代方法,称为类幂法,该方法定义明确。此外,我们还证明了类幂方法产生的特征值估计序列和特征向量评估序列分别线性收敛于谱半径及其相应的特征向量。为了加速该方法,我们引入了线搜索技术。改进后的方法保留了与原始版本相同的收敛特性。大量的数值结果表明,改进方法的性能相当出色。
{"title":"A power-like method for finding the spectral radius of a weakly irreducible nonnegative symmetric tensor","authors":"Xueli Bai, Dong-Hui Li, Lei Wu, Jiefeng Xu","doi":"10.1007/s10589-024-00601-8","DOIUrl":"https://doi.org/10.1007/s10589-024-00601-8","url":null,"abstract":"<p>The Perron–Frobenius theorem says that the spectral radius of a weakly irreducible nonnegative tensor is the unique positive eigenvalue corresponding to a positive eigenvector. With this fact in mind, the purpose of this paper is to find the spectral radius and its corresponding positive eigenvector of a weakly irreducible nonnegative symmetric tensor. By transforming the eigenvalue problem into an equivalent problem of minimizing a concave function on a closed convex set, we derive a simpler and cheaper iterative method called power-like method, which is well-defined. Furthermore, we show that both sequences of the eigenvalue estimates and the eigenvector evaluations generated by the power-like method <i>Q</i>-linearly converge to the spectral radius and its corresponding eigenvector, respectively. To accelerate the method, we introduce a line search technique. The improved method retains the same convergence property as the original version. Plentiful numerical results show that the improved method performs quite well.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"46 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142199606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An inexact regularized proximal Newton method without line search 无需直线搜索的非精确正则近似牛顿法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-08-16 DOI: 10.1007/s10589-024-00600-9
Simeon vom Dahl, Christian Kanzow

In this paper, we introduce an inexact regularized proximal Newton method (IRPNM) that does not require any line search. The method is designed to minimize the sum of a twice continuously differentiable function f and a convex (possibly non-smooth and extended-valued) function (varphi ). Instead of controlling a step size by a line search procedure, we update the regularization parameter in a suitable way, based on the success of the previous iteration. The global convergence of the sequence of iterations and its superlinear convergence rate under a local Hölderian error bound assumption are shown. Notably, these convergence results are obtained without requiring a global Lipschitz property for ( nabla f ), which, to the best of the authors’ knowledge, is a novel contribution for proximal Newton methods. To highlight the efficiency of our approach, we provide numerical comparisons with an IRPNM using a line search globalization and a modern FISTA-type method.

在本文中,我们介绍了一种不需要任何直线搜索的非精确正则化近似牛顿法(IRPNM)。该方法旨在最小化两次连续可微分函数 f 与凸(可能是非光滑和扩展值)函数 (varphi )之和。我们不是通过直线搜索程序来控制步长,而是根据前一次迭代的成功率,以适当的方式更新正则化参数。在局部霍尔德误差约束假设下,迭代序列的全局收敛性及其超线性收敛率得到了证明。值得注意的是,这些收敛结果是在不要求 ( nabla f ) 的全局 Lipschitz 属性的情况下获得的,据作者所知,这是近似牛顿方法的一个新贡献。为了突出我们方法的效率,我们提供了使用线搜索全局化的 IRPNM 和现代 FISTA 类型方法的数值比较。
{"title":"An inexact regularized proximal Newton method without line search","authors":"Simeon vom Dahl, Christian Kanzow","doi":"10.1007/s10589-024-00600-9","DOIUrl":"https://doi.org/10.1007/s10589-024-00600-9","url":null,"abstract":"<p>In this paper, we introduce an inexact regularized proximal Newton method (IRPNM) that does not require any line search. The method is designed to minimize the sum of a twice continuously differentiable function <i>f</i> and a convex (possibly non-smooth and extended-valued) function <span>(varphi )</span>. Instead of controlling a step size by a line search procedure, we update the regularization parameter in a suitable way, based on the success of the previous iteration. The global convergence of the sequence of iterations and its superlinear convergence rate under a local Hölderian error bound assumption are shown. Notably, these convergence results are obtained without requiring a global Lipschitz property for <span>( nabla f )</span>, which, to the best of the authors’ knowledge, is a novel contribution for proximal Newton methods. To highlight the efficiency of our approach, we provide numerical comparisons with an IRPNM using a line search globalization and a modern FISTA-type method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"29 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142199607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A block-coordinate approach of multi-level optimization with an application to physics-informed neural networks 应用于物理信息神经网络的多级优化块坐标方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-08-06 DOI: 10.1007/s10589-024-00597-1
Serge Gratton, Valentin Mercier, Elisa Riccietti, Philippe L. Toint

Multi-level methods are widely used for the solution of large-scale problems, because of their computational advantages and exploitation of the complementarity between the involved sub-problems. After a re-interpretation of multi-level methods from a block-coordinate point of view, we propose a multi-level algorithm for the solution of nonlinear optimization problems and analyze its evaluation complexity. We apply it to the solution of partial differential equations using physics-informed neural networks (PINNs) and consider two different types of neural architectures, a generic feedforward network and a frequency-aware network. We show that our approach is particularly effective if coupled with these specialized architectures and that this coupling results in better solutions and significant computational savings.

多层次方法因其计算优势和利用相关子问题之间的互补性而被广泛用于解决大规模问题。从块坐标的角度重新解释多层次方法后,我们提出了一种解决非线性优化问题的多层次算法,并分析了其评估复杂性。我们将其应用于使用物理信息神经网络(PINNs)求解偏微分方程,并考虑了两种不同类型的神经架构:普通前馈网络和频率感知网络。我们的研究表明,如果将我们的方法与这些专门的架构结合起来,会特别有效,而且这种结合会带来更好的解决方案,并显著节省计算量。
{"title":"A block-coordinate approach of multi-level optimization with an application to physics-informed neural networks","authors":"Serge Gratton, Valentin Mercier, Elisa Riccietti, Philippe L. Toint","doi":"10.1007/s10589-024-00597-1","DOIUrl":"https://doi.org/10.1007/s10589-024-00597-1","url":null,"abstract":"<p>Multi-level methods are widely used for the solution of large-scale problems, because of their computational advantages and exploitation of the complementarity between the involved sub-problems. After a re-interpretation of multi-level methods from a block-coordinate point of view, we propose a multi-level algorithm for the solution of nonlinear optimization problems and analyze its evaluation complexity. We apply it to the solution of partial differential equations using physics-informed neural networks (PINNs) and consider two different types of neural architectures, a generic feedforward network and a frequency-aware network. We show that our approach is particularly effective if coupled with these specialized architectures and that this coupling results in better solutions and significant computational savings.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"58 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141944946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Full-low evaluation methods for bound and linearly constrained derivative-free optimization 约束和线性约束无导数优化的全低评估方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-08-01 DOI: 10.1007/s10589-024-00596-2
C. W. Royer, O. Sohab, L. N. Vicente

Derivative-free optimization (DFO) consists in finding the best value of an objective function without relying on derivatives. To tackle such problems, one may build approximate derivatives, using for instance finite-difference estimates. One may also design algorithmic strategies that perform space exploration and seek improvement over the current point. The first type of strategy often provides good performance on smooth problems but at the expense of more function evaluations. The second type is cheaper and typically handles non-smoothness or noise in the objective better. Recently, full-low evaluation methods have been proposed as a hybrid class of DFO algorithms that combine both strategies, respectively denoted as Full-Eval and Low-Eval. In the unconstrained case, these methods showed promising numerical performance. In this paper, we extend the full-low evaluation framework to bound and linearly constrained derivative-free optimization. We derive convergence results for an instance of this framework, that combines finite-difference quasi-Newton steps with probabilistic direct-search steps. The former are projected onto the feasible set, while the latter are defined within tangent cones identified by nearby active constraints. We illustrate the practical performance of our instance on standard linearly constrained problems, that we adapt to introduce noisy evaluations as well as non-smoothness. In all cases, our method performs favorably compared to algorithms that rely solely on Full-eval or Low-eval iterations.

无导数优化(DFO)是指在不依赖导数的情况下找到目标函数的最佳值。为了解决这类问题,我们可以利用有限差分估计等方法建立近似导数。也可以设计算法策略,进行空间探索,寻求对当前点的改进。第一种策略通常能为平滑问题提供良好的性能,但代价是需要进行更多的函数评估。第二种策略成本更低,通常能更好地处理目标中的非平稳性或噪声。最近,有人提出了全低评估方法,作为 DFO 算法的混合类,将这两种策略结合起来,分别称为全评估和低评估。在无约束情况下,这些方法显示出良好的数值性能。在本文中,我们将全低评估框架扩展到有约束和线性约束的无导数优化。我们推导了该框架实例的收敛结果,该实例结合了有限差分准牛顿步骤和概率直接搜索步骤。前者投影到可行集上,而后者则定义在由附近主动约束确定的切线锥内。我们在标准线性约束问题上说明了我们的实例的实际性能,并对其进行了调整,以引入噪声评估和非平稳性。在所有情况下,我们的方法都优于仅依赖全值或低值迭代的算法。
{"title":"Full-low evaluation methods for bound and linearly constrained derivative-free optimization","authors":"C. W. Royer, O. Sohab, L. N. Vicente","doi":"10.1007/s10589-024-00596-2","DOIUrl":"https://doi.org/10.1007/s10589-024-00596-2","url":null,"abstract":"<p>Derivative-free optimization (DFO) consists in finding the best value of an objective function without relying on derivatives. To tackle such problems, one may build approximate derivatives, using for instance finite-difference estimates. One may also design algorithmic strategies that perform space exploration and seek improvement over the current point. The first type of strategy often provides good performance on smooth problems but at the expense of more function evaluations. The second type is cheaper and typically handles non-smoothness or noise in the objective better. Recently, full-low evaluation methods have been proposed as a hybrid class of DFO algorithms that combine both strategies, respectively denoted as Full-Eval and Low-Eval. In the unconstrained case, these methods showed promising numerical performance. In this paper, we extend the full-low evaluation framework to bound and linearly constrained derivative-free optimization. We derive convergence results for an instance of this framework, that combines finite-difference quasi-Newton steps with probabilistic direct-search steps. The former are projected onto the feasible set, while the latter are defined within tangent cones identified by nearby active constraints. We illustrate the practical performance of our instance on standard linearly constrained problems, that we adapt to introduce noisy evaluations as well as non-smoothness. In all cases, our method performs favorably compared to algorithms that rely solely on Full-eval or Low-eval iterations.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"295 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141880890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An adaptive regularized proximal Newton-type methods for composite optimization over the Stiefel manifold 用于斯特菲尔流形上复合优化的自适应正则近端牛顿型方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-07-26 DOI: 10.1007/s10589-024-00595-3
Qinsi Wang, Wei Hong Yang

Recently, the proximal Newton-type method and its variants have been generalized to solve composite optimization problems over the Stiefel manifold whose objective function is the summation of a smooth function and a nonsmooth function. In this paper, we propose an adaptive quadratically regularized proximal quasi-Newton method, named ARPQN, to solve this class of problems. Under some mild assumptions, the global convergence, the local linear convergence rate and the iteration complexity of ARPQN are established. Numerical experiments and comparisons with other state-of-the-art methods indicate that ARPQN is very promising. We also propose an adaptive quadratically regularized proximal Newton method, named ARPN. It is shown the ARPN method has a local superlinear convergence rate under certain reasonable assumptions, which demonstrates attractive convergence properties of regularized proximal Newton methods.

最近,近似牛顿法及其变体被推广用于求解目标函数为光滑函数与非光滑函数之和的 Stiefel 流形上的复合优化问题。本文提出了一种名为 ARPQN 的自适应二次正则近似准牛顿法来解决这类问题。在一些温和的假设条件下,建立了 ARPQN 的全局收敛性、局部线性收敛率和迭代复杂度。数值实验以及与其他最先进方法的比较表明,ARPQN 非常有前途。我们还提出了一种自适应二次正则化近牛顿方法,命名为 ARPN。结果表明,在某些合理的假设条件下,ARPN 方法具有局部超线性收敛率,这证明了正则化近牛顿方法具有诱人的收敛特性。
{"title":"An adaptive regularized proximal Newton-type methods for composite optimization over the Stiefel manifold","authors":"Qinsi Wang, Wei Hong Yang","doi":"10.1007/s10589-024-00595-3","DOIUrl":"https://doi.org/10.1007/s10589-024-00595-3","url":null,"abstract":"<p>Recently, the proximal Newton-type method and its variants have been generalized to solve composite optimization problems over the Stiefel manifold whose objective function is the summation of a smooth function and a nonsmooth function. In this paper, we propose an adaptive quadratically regularized proximal quasi-Newton method, named ARPQN, to solve this class of problems. Under some mild assumptions, the global convergence, the local linear convergence rate and the iteration complexity of ARPQN are established. Numerical experiments and comparisons with other state-of-the-art methods indicate that ARPQN is very promising. We also propose an adaptive quadratically regularized proximal Newton method, named ARPN. It is shown the ARPN method has a local superlinear convergence rate under certain reasonable assumptions, which demonstrates attractive convergence properties of regularized proximal Newton methods.\u0000</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"154 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141780194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computational Optimization and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1