首页 > 最新文献

Computational Optimization and Applications最新文献

英文 中文
Delayed Weighted Gradient Method with simultaneous step-sizes for strongly convex optimization 采用同步步长的延迟加权梯度法进行强凸优化
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-05-31 DOI: 10.1007/s10589-024-00586-4
Hugo Lara, Rafael Aleixo, Harry Oviedo

The Delayed Weighted Gradient Method (DWGM) is a two-step gradient algorithm that is efficient for the minimization of large scale strictly convex quadratic functions. It has orthogonality properties that make it to compete with the Conjugate Gradient (CG) method. Both methods calculate in sequence two step-sizes, CG minimizes the objective function and DWGM the gradient norm, alongside two search directions defined over first order current and previous iteration information. The objective of this work is to accelerate the recently developed extension of DWGM to nonquadratic strongly convex minimization problems. Our idea is to define the step-sizes of DWGM in a unique two dimensional convex quadratic optimization problem, calculating them simultaneously. Convergence of the resulting algorithm is analyzed. Comparative numerical experiments illustrate the effectiveness of our approach.

延迟加权梯度法(DWGM)是一种两步梯度算法,对于大规模严格凸二次函数的最小化非常有效。它具有正交特性,可与共轭梯度法(CG)相媲美。这两种方法都依次计算两个步长,CG 最小化目标函数,DWGM 最小化梯度规范,同时根据一阶当前和前一次迭代信息定义两个搜索方向。这项工作的目的是加速最近开发的 DWGM 对非二次强凸最小化问题的扩展。我们的想法是在一个独特的二维凸二次优化问题中定义 DWGM 的步长,同时计算它们。我们分析了算法的收敛性。对比数值实验说明了我们方法的有效性。
{"title":"Delayed Weighted Gradient Method with simultaneous step-sizes for strongly convex optimization","authors":"Hugo Lara, Rafael Aleixo, Harry Oviedo","doi":"10.1007/s10589-024-00586-4","DOIUrl":"https://doi.org/10.1007/s10589-024-00586-4","url":null,"abstract":"<p>The Delayed Weighted Gradient Method (DWGM) is a two-step gradient algorithm that is efficient for the minimization of large scale strictly convex quadratic functions. It has orthogonality properties that make it to compete with the Conjugate Gradient (CG) method. Both methods calculate in sequence two step-sizes, CG minimizes the objective function and DWGM the gradient norm, alongside two search directions defined over first order current and previous iteration information. The objective of this work is to accelerate the recently developed extension of DWGM to nonquadratic strongly convex minimization problems. Our idea is to define the step-sizes of DWGM in a unique two dimensional convex quadratic optimization problem, calculating them simultaneously. Convergence of the resulting algorithm is analyzed. Comparative numerical experiments illustrate the effectiveness of our approach.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"2012 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141195263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A non-monotone trust-region method with noisy oracles and additional sampling 具有噪声信标和额外采样的非单调信任区域方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-05-31 DOI: 10.1007/s10589-024-00580-w
Nataša Krejić, Nataša Krklec Jerinkić, Ángeles Martínez, Mahsa Yousefi

In this work, we introduce a novel stochastic second-order method, within the framework of a non-monotone trust-region approach, for solving the unconstrained, nonlinear, and non-convex optimization problems arising in the training of deep neural networks. The proposed algorithm makes use of subsampling strategies that yield noisy approximations of the finite sum objective function and its gradient. We introduce an adaptive sample size strategy based on inexpensive additional sampling to control the resulting approximation error. Depending on the estimated progress of the algorithm, this can yield sample size scenarios ranging from mini-batch to full sample functions. We provide convergence analysis for all possible scenarios and show that the proposed method achieves almost sure convergence under standard assumptions for the trust-region framework. We report numerical experiments showing that the proposed algorithm outperforms its state-of-the-art counterpart in deep neural network training for image classification and regression tasks while requiring a significantly smaller number of gradient evaluations.

在这项工作中,我们在非单调信任区域方法的框架内引入了一种新型随机二阶方法,用于解决深度神经网络训练中出现的无约束、非线性和非凸优化问题。所提出的算法采用了子采样策略,可以得到有限和目标函数及其梯度的噪声近似值。我们引入了一种基于廉价额外采样的自适应样本大小策略,以控制由此产生的近似误差。根据算法的估计进度,这可以产生从小批量到全样本函数的样本大小方案。我们提供了所有可能方案的收敛性分析,并表明在信任区域框架的标准假设条件下,所提出的方法几乎可以确保收敛性。我们报告的数值实验表明,在针对图像分类和回归任务的深度神经网络训练中,所提出的算法优于其最先进的同类算法,同时所需的梯度评估次数也大大减少。
{"title":"A non-monotone trust-region method with noisy oracles and additional sampling","authors":"Nataša Krejić, Nataša Krklec Jerinkić, Ángeles Martínez, Mahsa Yousefi","doi":"10.1007/s10589-024-00580-w","DOIUrl":"https://doi.org/10.1007/s10589-024-00580-w","url":null,"abstract":"<p>In this work, we introduce a novel stochastic second-order method, within the framework of a non-monotone trust-region approach, for solving the unconstrained, nonlinear, and non-convex optimization problems arising in the training of deep neural networks. The proposed algorithm makes use of subsampling strategies that yield noisy approximations of the finite sum objective function and its gradient. We introduce an adaptive sample size strategy based on inexpensive additional sampling to control the resulting approximation error. Depending on the estimated progress of the algorithm, this can yield sample size scenarios ranging from mini-batch to full sample functions. We provide convergence analysis for all possible scenarios and show that the proposed method achieves almost sure convergence under standard assumptions for the trust-region framework. We report numerical experiments showing that the proposed algorithm outperforms its state-of-the-art counterpart in deep neural network training for image classification and regression tasks while requiring a significantly smaller number of gradient evaluations.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"87 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141195413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A boosted DC algorithm for non-differentiable DC components with non-monotone line search 采用非单调线搜索的无差别直流分量提升直流算法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-05-11 DOI: 10.1007/s10589-024-00578-4
O. P. Ferreira, E. M. Santos, J. C. O. Souza

We introduce a new approach to apply the boosted difference of convex functions algorithm (BDCA) for solving non-convex and non-differentiable problems involving difference of two convex functions (DC functions). Supposing the first DC component differentiable and the second one possibly non-differentiable, the main idea of BDCA is to use the point computed by the subproblem of the DC algorithm (DCA) to define a descent direction of the objective from that point, and then a monotone line search starting from it is performed in order to find a new point which decreases the objective function when compared with the point generated by the subproblem of DCA. This procedure improves the performance of the DCA. However, if the first DC component is non-differentiable, then the direction computed by BDCA can be an ascent direction and a monotone line search cannot be performed. Our approach uses a non-monotone line search in the BDCA (nmBDCA) to enable a possible growth in the objective function values controlled by a parameter. Under suitable assumptions, we show that any cluster point of the sequence generated by the nmBDCA is a critical point of the problem under consideration and provides some iteration-complexity bounds. Furthermore, if the first DC component is differentiable, we present different iteration-complexity bounds and prove the full convergence of the sequence under the Kurdyka–Łojasiewicz property of the objective function. Some numerical experiments show that the nmBDCA outperforms the DCA, such as its monotone version.

我们引入了一种新方法,将凸函数差分算法(BDCA)应用于解决涉及两个凸函数(DC 函数)差分的非凸和非微分问题。假设第一个凸函数分量是可微分的,而第二个分量可能是不可微分的,BDCA 的主要思想是利用凸函数算法(DCA)子问题计算出的点来定义从该点开始的目标下降方向,然后从该点开始进行单调直线搜索,以找到一个与 DCA 子问题产生的点相比目标函数减小的新点。这一过程提高了 DCA 的性能。但是,如果第一个直流分量是无差别的,那么 BDCA 计算出的方向可能是一个上升方向,无法进行单调线搜索。我们的方法在 BDCA 中使用非单调线性搜索(nmBDCA),使目标函数值的增长可能受参数控制。在适当的假设条件下,我们证明了 nmBDCA 生成的序列中的任何簇点都是所考虑问题的临界点,并提供了一些迭代复杂度边界。此外,如果第一个直流分量是可微分的,我们提出了不同的迭代复杂度边界,并证明了在目标函数的 Kurdyka-Łojasiewicz 特性下序列的完全收敛性。一些数值实验表明,nmBDCA 优于 DCA(如其单调版本)。
{"title":"A boosted DC algorithm for non-differentiable DC components with non-monotone line search","authors":"O. P. Ferreira, E. M. Santos, J. C. O. Souza","doi":"10.1007/s10589-024-00578-4","DOIUrl":"https://doi.org/10.1007/s10589-024-00578-4","url":null,"abstract":"<p>We introduce a new approach to apply the boosted difference of convex functions algorithm (BDCA) for solving non-convex and non-differentiable problems involving difference of two convex functions (DC functions). Supposing the first DC component differentiable and the second one possibly non-differentiable, the main idea of BDCA is to use the point computed by the subproblem of the DC algorithm (DCA) to define a descent direction of the objective from that point, and then a monotone line search starting from it is performed in order to find a new point which decreases the objective function when compared with the point generated by the subproblem of DCA. This procedure improves the performance of the DCA. However, if the first DC component is non-differentiable, then the direction computed by BDCA can be an ascent direction and a monotone line search cannot be performed. Our approach uses a non-monotone line search in the BDCA (nmBDCA) to enable a possible growth in the objective function values controlled by a parameter. Under suitable assumptions, we show that any cluster point of the sequence generated by the nmBDCA is a critical point of the problem under consideration and provides some iteration-complexity bounds. Furthermore, if the first DC component is differentiable, we present different iteration-complexity bounds and prove the full convergence of the sequence under the Kurdyka–Łojasiewicz property of the objective function. Some numerical experiments show that the nmBDCA outperforms the DCA, such as its monotone version.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"23 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An away-step Frank–Wolfe algorithm for constrained multiobjective optimization 用于约束性多目标优化的远离步骤弗兰克-沃尔夫算法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-05-07 DOI: 10.1007/s10589-024-00577-5
Douglas S. Gonçalves, Max L. N. Gonçalves, Jefferson G. Melo

In this paper, we propose and analyze an away-step Frank–Wolfe algorithm designed for solving multiobjective optimization problems over polytopes. We prove that each limit point of the sequence generated by the algorithm is a weak Pareto optimal solution. Furthermore, under additional conditions, we show linear convergence of the whole sequence to a Pareto optimal solution. Numerical examples illustrate a promising performance of the proposed algorithm in problems where the multiobjective Frank–Wolfe convergence rate is only sublinear.

在本文中,我们提出并分析了一种专为解决多边形上的多目标优化问题而设计的分步 Frank-Wolfe 算法。我们证明了该算法生成的序列的每个极限点都是弱帕累托最优解。此外,在附加条件下,我们还证明了整个序列对帕累托最优解的线性收敛性。数值示例表明,在多目标 Frank-Wolfe 收敛率仅为亚线性的问题中,所提出的算法表现出良好的性能。
{"title":"An away-step Frank–Wolfe algorithm for constrained multiobjective optimization","authors":"Douglas S. Gonçalves, Max L. N. Gonçalves, Jefferson G. Melo","doi":"10.1007/s10589-024-00577-5","DOIUrl":"https://doi.org/10.1007/s10589-024-00577-5","url":null,"abstract":"<p>In this paper, we propose and analyze an <i>away-step</i> Frank–Wolfe algorithm designed for solving multiobjective optimization problems over polytopes. We prove that each limit point of the sequence generated by the algorithm is a weak Pareto optimal solution. Furthermore, under additional conditions, we show linear convergence of the whole sequence to a Pareto optimal solution. Numerical examples illustrate a promising performance of the proposed algorithm in problems where the multiobjective Frank–Wolfe convergence rate is only sublinear.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"65 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140887560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shape optimization for interface identification in nonlocal models 非局部模型界面识别的形状优化
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-05-07 DOI: 10.1007/s10589-024-00575-7
Matthias Schuster, Christian Vollmann, Volker Schulz

Shape optimization methods have been proven useful for identifying interfaces in models governed by partial differential equations. Here we consider a class of shape optimization problems constrained by nonlocal equations which involve interface–dependent kernels. We derive a novel shape derivative associated to the nonlocal system model and solve the problem by established numerical techniques. The code for obtaining the results in this paper is published at (https://github.com/schustermatthias/nlshape).

事实证明,形状优化方法有助于识别偏微分方程模型中的界面。在此,我们考虑一类受非局部方程约束的形状优化问题,其中涉及与界面相关的核。我们推导出一种与非局部系统模型相关的新型形状导数,并通过成熟的数值技术解决该问题。获得本文结果的代码发布于 (https://github.com/schustermatthias/nlshape)。
{"title":"Shape optimization for interface identification in nonlocal models","authors":"Matthias Schuster, Christian Vollmann, Volker Schulz","doi":"10.1007/s10589-024-00575-7","DOIUrl":"https://doi.org/10.1007/s10589-024-00575-7","url":null,"abstract":"<p>Shape optimization methods have been proven useful for identifying interfaces in models governed by partial differential equations. Here we consider a class of shape optimization problems constrained by nonlocal equations which involve interface–dependent kernels. We derive a novel shape derivative associated to the nonlocal system model and solve the problem by established numerical techniques. The code for obtaining the results in this paper is published at (https://github.com/schustermatthias/nlshape).</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"204 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hybrid inexact regularized Newton and negative curvature method 不精确正则化牛顿和负曲率混合方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-05-06 DOI: 10.1007/s10589-024-00576-6
Hong Zhu, Yunhai Xiao

In this paper, we propose a hybrid inexact regularized Newton and negative curvature method for solving unconstrained nonconvex problems. The descent direction is chosen based on different conditions, either the negative curvature or the inexact regularized direction. In addition, to minimize computational costs while obtaining the negative curvature, we employ a dimensionality reduction strategy to verify if the Hessian matrix exhibits negative curvatures within a three-dimensional subspace. We show that the proposed method can achieve the best-known global iteration complexity if the Hessian of the objective function is Lipschitz continuous on a certain compact set. Two simplified methods for nonconvex and strongly convex problems are analyzed as specific instances of the proposed method. We show that under the local error bound assumption with respect to the gradient, the distance between iterations generated by our proposed method and the local solution set converges to (0) at a superlinear rate. Additionally, for strongly convex problems, the quadratic convergence rate can be achieved. Extensive numerical experiments show the effectiveness of the proposed method.

本文提出了一种非精确正则牛顿和负曲率混合方法,用于解决无约束非凸问题。根据不同的条件选择下降方向,可以是负曲率方向,也可以是非精确正则化方向。此外,为了在获得负曲率的同时最大限度地降低计算成本,我们采用了降维策略,以验证 Hessian 矩阵是否在三维子空间内呈现负曲率。我们的研究表明,如果目标函数的 Hessian 在某个紧凑集合上是 Lipschitz 连续的,那么所提出的方法就能达到已知的最佳全局迭代复杂度。作为所提方法的具体实例,我们分析了针对非凸问题和强凸问题的两种简化方法。我们证明,在梯度的局部误差约束假设下,我们提出的方法产生的迭代与局部解集之间的距离以超线性速率收敛到(0)。此外,对于强凸问题,可以达到二次收敛率。大量的数值实验表明了所提方法的有效性。
{"title":"A hybrid inexact regularized Newton and negative curvature method","authors":"Hong Zhu, Yunhai Xiao","doi":"10.1007/s10589-024-00576-6","DOIUrl":"https://doi.org/10.1007/s10589-024-00576-6","url":null,"abstract":"<p>In this paper, we propose a hybrid inexact regularized Newton and negative curvature method for solving unconstrained nonconvex problems. The descent direction is chosen based on different conditions, either the negative curvature or the inexact regularized direction. In addition, to minimize computational costs while obtaining the negative curvature, we employ a dimensionality reduction strategy to verify if the Hessian matrix exhibits negative curvatures within a three-dimensional subspace. We show that the proposed method can achieve the best-known global iteration complexity if the Hessian of the objective function is Lipschitz continuous on a certain compact set. Two simplified methods for nonconvex and strongly convex problems are analyzed as specific instances of the proposed method. We show that under the local error bound assumption with respect to the gradient, the distance between iterations generated by our proposed method and the local solution set converges to <span>(0)</span> at a superlinear rate. Additionally, for strongly convex problems, the quadratic convergence rate can be achieved. Extensive numerical experiments show the effectiveness of the proposed method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"24 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140887643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chance-constrained programs with convex underlying functions: a bilevel convex optimization perspective 具有凸基础函数的机会受限程序:双层凸优化视角
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-04-27 DOI: 10.1007/s10589-024-00573-9
Yassine Laguel, Jérôme Malick, Wim van Ackooij

Chance constraints are a valuable tool for the design of safe decisions in uncertain environments; they are used to model satisfaction of a constraint with a target probability. However, because of possible non-convexity and non-smoothness, optimizing over a chance constrained set is challenging. In this paper, we consider chance constrained programs where the objective function and the constraints are convex with respect to the decision parameter. We establish an exact reformulation of such a problem as a bilevel problem with a convex lower-level. Then we leverage this bilevel formulation to propose a tractable penalty approach, in the setting of finitely supported random variables. The penalized objective is a difference-of-convex function that we minimize with a suitable bundle algorithm. We release an easy-to-use open-source python toolbox implementing the approach, with a special emphasis on fast computational subroutines.

偶然性约束是在不确定环境中设计安全决策的重要工具;偶然性约束用于模拟满足目标概率的约束。然而,由于可能存在非凸性和非光滑性,在偶然约束集上进行优化具有挑战性。在本文中,我们考虑了目标函数和约束条件相对于决策参数都是凸的偶然约束程序。我们将此类问题精确地重新表述为具有凸低级问题的双级问题。然后,我们在有限支持随机变量的背景下,利用这种双层表述,提出了一种可行的惩罚方法。受惩罚的目标是一个凸函数差,我们用合适的捆绑算法将其最小化。我们发布了一个易于使用的开源 python 工具箱来实现这种方法,并特别强调了快速计算子程序。
{"title":"Chance-constrained programs with convex underlying functions: a bilevel convex optimization perspective","authors":"Yassine Laguel, Jérôme Malick, Wim van Ackooij","doi":"10.1007/s10589-024-00573-9","DOIUrl":"https://doi.org/10.1007/s10589-024-00573-9","url":null,"abstract":"<p>Chance constraints are a valuable tool for the design of safe decisions in uncertain environments; they are used to model satisfaction of a constraint with a target probability. However, because of possible non-convexity and non-smoothness, optimizing over a chance constrained set is challenging. In this paper, we consider chance constrained programs where the objective function and the constraints are convex with respect to the decision parameter. We establish an exact reformulation of such a problem as a bilevel problem with a convex lower-level. Then we leverage this bilevel formulation to propose a tractable penalty approach, in the setting of finitely supported random variables. The penalized objective is a difference-of-convex function that we minimize with a suitable bundle algorithm. We release an easy-to-use open-source python toolbox implementing the approach, with a special emphasis on fast computational subroutines.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"20 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140809506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic average model methods 随机平均模型方法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-04-24 DOI: 10.1007/s10589-024-00563-x
Matt Menickelly, Stefan M. Wild

We consider the solution of finite-sum minimization problems, such as those appearing in nonlinear least-squares or general empirical risk minimization problems. We are motivated by problems in which the summand functions are computationally expensive and evaluating all summands on every iteration of an optimization method may be undesirable. We present the idea of stochastic average model (SAM) methods, inspired by stochastic average gradient methods. SAM methods sample component functions on each iteration of a trust-region method according to a discrete probability distribution on component functions; the distribution is designed to minimize an upper bound on the variance of the resulting stochastic model. We present promising numerical results concerning an implemented variant extending the derivative-free model-based trust-region solver POUNDERS, which we name SAM-POUNDERS.

我们考虑求解有限求和最小化问题,例如非线性最小二乘法或一般经验风险最小化问题中出现的问题。我们考虑的问题是,求和函数的计算成本很高,而且在优化方法的每次迭代中评估所有求和函数可能并不可取。受随机平均梯度法的启发,我们提出了随机平均模型(SAM)方法。SAM 方法根据分量函数的离散概率分布,在信任区域方法的每次迭代中对分量函数进行采样;该分布旨在最小化随机模型方差的上限。我们介绍了扩展基于模型的无导数信任区域求解器 POUNDERS 的实施变体,并将其命名为 SAM-POUNDERS,该变体的数值结果很有希望。
{"title":"Stochastic average model methods","authors":"Matt Menickelly, Stefan M. Wild","doi":"10.1007/s10589-024-00563-x","DOIUrl":"https://doi.org/10.1007/s10589-024-00563-x","url":null,"abstract":"<p>We consider the solution of finite-sum minimization problems, such as those appearing in nonlinear least-squares or general empirical risk minimization problems. We are motivated by problems in which the summand functions are computationally expensive and evaluating all summands on every iteration of an optimization method may be undesirable. We present the idea of stochastic average model (<span>SAM</span>) methods, inspired by stochastic average gradient methods. <span>SAM</span> methods sample component functions on each iteration of a trust-region method according to a discrete probability distribution on component functions; the distribution is designed to minimize an upper bound on the variance of the resulting stochastic model. We present promising numerical results concerning an implemented variant extending the derivative-free model-based trust-region solver <span>POUNDERS</span>, which we name <span>SAM-POUNDERS</span>.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"42 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140804672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Safeguarded augmented Lagrangian algorithms with scaled stopping criterion for the subproblems 对子问题采用按比例停止标准的有保障的扩增拉格朗日算法
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-04-15 DOI: 10.1007/s10589-024-00572-w
E. G. Birgin, G. Haeser, J. M. Martínez

At each iteration of the safeguarded augmented Lagrangian algorithm Algencan, a bound-constrained subproblem consisting of the minimization of the Powell–Hestenes–Rockafellar augmented Lagrangian function is considered, for which an approximate minimizer with tolerance tending to zero is sought. More precisely, a point that satisfies a subproblem first-order necessary optimality condition with tolerance tending to zero is required. In this work, based on the success of scaled stopping criteria in constrained optimization, we propose a scaled stopping criterion for the subproblems of Algencan. The scaling is done with the maximum absolute value of the first-order Lagrange multipliers approximation, whenever it is larger than one. The difference between the convergence theory of the scaled and non-scaled versions of Algencan is discussed and extensive numerical experiments are provided.

在保障性扩增拉格朗日算法 Algencan 的每次迭代中,都要考虑一个有约束的子问题,即 Powell-Hestenes-Rockafellar 扩增拉格朗日函数的最小化问题,并为该问题寻找一个容差趋于零的近似最小值。更确切地说,需要一个满足子问题一阶必要最优条件且容差趋于零的点。在这项工作中,基于约束优化中按比例停止准则的成功经验,我们为 Algencan 的子问题提出了一种按比例停止准则。只要一阶拉格朗日乘数近似值的最大绝对值大于 1,就会按比例停止。本文讨论了 Algencan 的缩放和非缩放版本的收敛理论之间的差异,并提供了大量的数值实验。
{"title":"Safeguarded augmented Lagrangian algorithms with scaled stopping criterion for the subproblems","authors":"E. G. Birgin, G. Haeser, J. M. Martínez","doi":"10.1007/s10589-024-00572-w","DOIUrl":"https://doi.org/10.1007/s10589-024-00572-w","url":null,"abstract":"<p>At each iteration of the safeguarded augmented Lagrangian algorithm Algencan, a bound-constrained subproblem consisting of the minimization of the Powell–Hestenes–Rockafellar augmented Lagrangian function is considered, for which an approximate minimizer with tolerance tending to zero is sought. More precisely, a point that satisfies a subproblem first-order necessary optimality condition with tolerance tending to zero is required. In this work, based on the success of scaled stopping criteria in constrained optimization, we propose a scaled stopping criterion for the subproblems of Algencan. The scaling is done with the maximum absolute value of the first-order Lagrange multipliers approximation, whenever it is larger than one. The difference between the convergence theory of the scaled and non-scaled versions of Algencan is discussed and extensive numerical experiments are provided.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"118 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140578554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Global convergence of a BFGS-type algorithm for nonconvex multiobjective optimization problems 非凸多目标优化问题的 BFGS 型算法的全局收敛性
IF 2.2 2区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2024-04-11 DOI: 10.1007/s10589-024-00571-x
L. F. Prudente, D. R. Souza

We propose a modified BFGS algorithm for multiobjective optimization problems with global convergence, even in the absence of convexity assumptions on the objective functions. Furthermore, we establish a local superlinear rate of convergence of the method under usual conditions. Our approach employs Wolfe step sizes and ensures that the Hessian approximations are updated and corrected at each iteration to address the lack of convexity assumption. Numerical results shows that the introduced modifications preserve the practical efficiency of the BFGS method.

我们针对多目标优化问题提出了一种改进的 BFGS 算法,该算法即使在目标函数不存在凸性假设的情况下也具有全局收敛性。此外,我们还确定了该方法在通常条件下的局部超线性收敛率。我们的方法采用了沃尔夫步长,并确保在每次迭代时更新和修正赫塞斯近似值,以解决缺乏凸性假设的问题。数值结果表明,引入的修改保持了 BFGS 方法的实用效率。
{"title":"Global convergence of a BFGS-type algorithm for nonconvex multiobjective optimization problems","authors":"L. F. Prudente, D. R. Souza","doi":"10.1007/s10589-024-00571-x","DOIUrl":"https://doi.org/10.1007/s10589-024-00571-x","url":null,"abstract":"<p>We propose a modified BFGS algorithm for multiobjective optimization problems with global convergence, even in the absence of convexity assumptions on the objective functions. Furthermore, we establish a local superlinear rate of convergence of the method under usual conditions. Our approach employs Wolfe step sizes and ensures that the Hessian approximations are updated and corrected at each iteration to address the lack of convexity assumption. Numerical results shows that the introduced modifications preserve the practical efficiency of the BFGS method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":"11 1","pages":""},"PeriodicalIF":2.2,"publicationDate":"2024-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140578780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computational Optimization and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1