Pub Date : 2024-01-22DOI: 10.1007/s10898-023-01335-3
Cao Thanh Tinh, Thai Doan Chuong
This paper studies a class of multiobjective convex polynomial problems, where both the constraint and objective functions involve uncertain parameters that reside in ellipsoidal uncertainty sets. Employing the robust deterministic approach, we provide necessary conditions and sufficient conditions, which are exhibited in relation to second order cone conditions, for robust (weak) Pareto solutions of the uncertain multiobjective optimization problem. A dual multiobjective problem is proposed to examine robust converse, robust weak and robust strong duality relations between the primal and dual problems. Moreover, we establish robust solution relationships between the uncertain multiobjective optimization program and a (scalar) second order cone programming relaxation problem of a corresponding weighted-sum optimization problem. This in particular shows that we can find a robust (weak) Pareto solution of the uncertain multiobjective optimization problem by solving a second order cone programming relaxation.
{"title":"Robust second order cone conditions and duality for multiobjective problems under uncertainty data","authors":"Cao Thanh Tinh, Thai Doan Chuong","doi":"10.1007/s10898-023-01335-3","DOIUrl":"https://doi.org/10.1007/s10898-023-01335-3","url":null,"abstract":"<p>This paper studies a class of multiobjective convex polynomial problems, where both the constraint and objective functions involve uncertain parameters that reside in ellipsoidal uncertainty sets. Employing the robust deterministic approach, we provide necessary conditions and sufficient conditions, which are exhibited in relation to second order cone conditions, for robust (weak) Pareto solutions of the uncertain multiobjective optimization problem. A dual multiobjective problem is proposed to examine robust converse, robust weak and robust strong duality relations between the primal and dual problems. Moreover, we establish robust solution relationships between the uncertain multiobjective optimization program and a (scalar) second order cone programming relaxation problem of a corresponding weighted-sum optimization problem. This in particular shows that we can find a robust (weak) Pareto solution of the uncertain multiobjective optimization problem by solving a second order cone programming relaxation.\u0000</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"1 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139518645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10898-023-01354-0
Sanchit Singh, Subhash C. Sarin, Ming Cheng
In this paper, we consider an application of lot-streaming for processing a lot of multiple items in a hybrid flow shop (HFS) for the objective of minimizing makespan. The HFS that we consider consists of two stages with a single machine available for processing in Stage 1 and m identical parallel machines in Stage 2. We call this problem a 1 + m TSHFS-LSP (two-stage hybrid flow shop, lot streaming problem), and show it to be NP-hard in general, except for the case when the sublot sizes are treated to be continuous. The novelty of our work is in obtaining closed-form expressions for optimal continuous sublot sizes that can be solved in polynomial time, for a given number of sublots. A fast linear search algorithm is also developed for determining the optimal number of sublots for the case of continuous sublot sizes. For the case when the sublot sizes are discrete, we propose a branch-and-bound-based heuristic to determine both the number of sublots and sublot sizes and demonstrate its efficacy by comparing its performance against that of a direct solution of a mixed-integer formulation of the problem by CPLEX®.
在本文中,我们考虑在混合流程车间(HFS)中应用批量流来处理包含多个项目的批量,以实现最小生产间隔的目标。我们考虑的 HFS 由两个阶段组成,第一阶段有一台可用于加工的机器,第二阶段有 m 台相同的并行机器。我们把这个问题称为 1 + m TSHFS-LSP(两阶段混合流程车间,批量流问题),并证明它在一般情况下是 NP-困难的,但子批量大小被视为连续的情况除外。我们工作的新颖之处在于获得了最优连续子批量大小的闭式表达式,对于给定数量的子批量,可以在多项式时间内求解。我们还开发了一种快速线性搜索算法,用于确定连续子槽尺寸情况下的最佳子槽数量。对于子槽大小离散的情况,我们提出了一种基于分支和边界的启发式方法来确定子槽数量和子槽大小,并通过比较其性能与 CPLEX® 对问题的混合整数表述的直接求解性能来证明其有效性。
{"title":"Single-lot, lot-streaming problem for a 1 + m hybrid flow shop","authors":"Sanchit Singh, Subhash C. Sarin, Ming Cheng","doi":"10.1007/s10898-023-01354-0","DOIUrl":"https://doi.org/10.1007/s10898-023-01354-0","url":null,"abstract":"<p>In this paper, we consider an application of lot-streaming for processing a lot of multiple items in a hybrid flow shop (HFS) for the objective of minimizing makespan. The HFS that we consider consists of two stages with a single machine available for processing in Stage 1 and <i>m</i> identical parallel machines in Stage 2. We call this problem a 1 + <i>m</i> TSHFS-LSP (two-stage hybrid flow shop, lot streaming problem), and show it to be NP-hard in general, except for the case when the sublot sizes are treated to be continuous. The novelty of our work is in obtaining closed-form expressions for optimal continuous sublot sizes that can be solved in polynomial time, for a given number of sublots. A fast linear search algorithm is also developed for determining the optimal number of sublots for the case of continuous sublot sizes. For the case when the sublot sizes are discrete, we propose a branch-and-bound-based heuristic to determine both the number of sublots and sublot sizes and demonstrate its efficacy by comparing its performance against that of a direct solution of a mixed-integer formulation of the problem by CPLEX<sup>®</sup>.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"2 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139458859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-06DOI: 10.1007/s10898-023-01353-1
Xiangkai Sun, Jiayi Huang, Kok Lay Teo
In this paper, we deal with a new class of SOS-convex (sum of squares convex) polynomial optimization problems with spectrahedral uncertainty data in both the objective and constraints. By using robust optimization and a weighted-sum scalarization methodology, we first present the relationship between robust solutions of this uncertain SOS-convex polynomial optimization problem and that of its corresponding scalar optimization problem. Then, by using a normal cone constraint qualification condition, we establish necessary and sufficient optimality conditions for robust weakly efficient solutions of this uncertain SOS-convex polynomial optimization problem based on scaled diagonally dominant sums of squares conditions and linear matrix inequalities. Moreover, we introduce a semidefinite programming relaxation problem of its weighted-sum scalar optimization problem, and show that robust weakly efficient solutions of the uncertain SOS-convex polynomial optimization problem can be found by solving the corresponding semidefinite programming relaxation problem.
本文讨论了一类新的 SOS-凸(平方凸和)多项式优化问题,该问题的目标和约束条件中均包含光谱不确定性数据。通过使用鲁棒优化和加权求和标量化方法,我们首先提出了这种不确定 SOS 凸多项式优化问题的鲁棒解与其相应标量优化问题的鲁棒解之间的关系。然后,通过使用法锥约束限定条件,我们基于比例对角显性平方和条件和线性矩阵不等式,建立了该不确定 SOS 凸多项式优化问题的鲁棒弱有效解的必要和充分最优条件。此外,我们还引入了其加权和标量优化问题的半有限编程松弛问题,并证明通过求解相应的半有限编程松弛问题,可以找到不确定 SOS 凸多项式优化问题的稳健弱高效解。
{"title":"On semidefinite programming relaxations for a class of robust SOS-convex polynomial optimization problems","authors":"Xiangkai Sun, Jiayi Huang, Kok Lay Teo","doi":"10.1007/s10898-023-01353-1","DOIUrl":"https://doi.org/10.1007/s10898-023-01353-1","url":null,"abstract":"<p>In this paper, we deal with a new class of SOS-convex (sum of squares convex) polynomial optimization problems with spectrahedral uncertainty data in both the objective and constraints. By using robust optimization and a weighted-sum scalarization methodology, we first present the relationship between robust solutions of this uncertain SOS-convex polynomial optimization problem and that of its corresponding scalar optimization problem. Then, by using a normal cone constraint qualification condition, we establish necessary and sufficient optimality conditions for robust weakly efficient solutions of this uncertain SOS-convex polynomial optimization problem based on scaled diagonally dominant sums of squares conditions and linear matrix inequalities. Moreover, we introduce a semidefinite programming relaxation problem of its weighted-sum scalar optimization problem, and show that robust weakly efficient solutions of the uncertain SOS-convex polynomial optimization problem can be found by solving the corresponding semidefinite programming relaxation problem.\u0000</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"206 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139373847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-05DOI: 10.1007/s10898-023-01349-x
Jing-jing Wang, Li-ping Tang, Xin-min Yang
In this paper, we propose a spectral projected subgradient method with a 1-memory momentum term for solving constrained convex multiobjective optimization problem. This method combines the subgradient-type algorithm for multiobjective optimization problems with the idea of the spectral projected algorithm to accelerate the convergence process. Additionally, a 1-memory momentum term is added to the subgradient direction in the early iterations. The 1-memory momentum term incorporates, in the present iteration, some of the influence of the past iterations, and this can help to improve the search direction. Under suitable assumptions, we show that the sequence generated by the method converges to a weakly Pareto efficient solution and derive the sublinear convergence rates for the proposed method. Finally, computational experiments are given to demonstrate the effectiveness of the proposed method.
{"title":"Spectral projected subgradient method with a 1-memory momentum term for constrained multiobjective optimization problem","authors":"Jing-jing Wang, Li-ping Tang, Xin-min Yang","doi":"10.1007/s10898-023-01349-x","DOIUrl":"https://doi.org/10.1007/s10898-023-01349-x","url":null,"abstract":"<p>In this paper, we propose a spectral projected subgradient method with a 1-memory momentum term for solving constrained convex multiobjective optimization problem. This method combines the subgradient-type algorithm for multiobjective optimization problems with the idea of the spectral projected algorithm to accelerate the convergence process. Additionally, a 1-memory momentum term is added to the subgradient direction in the early iterations. The 1-memory momentum term incorporates, in the present iteration, some of the influence of the past iterations, and this can help to improve the search direction. Under suitable assumptions, we show that the sequence generated by the method converges to a weakly Pareto efficient solution and derive the sublinear convergence rates for the proposed method. Finally, computational experiments are given to demonstrate the effectiveness of the proposed method.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"31 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139373808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-03DOI: 10.1007/s10898-023-01350-4
Abstract
A new exact projective penalty method is proposed for the equivalent reduction of constrained optimization problems to nonsmooth unconstrained ones. In the method, the original objective function is extended to infeasible points by summing its value at the projection of an infeasible point on the feasible set with the distance to the projection. Beside Euclidean projections, also a pointed projection in the direction of some fixed internal feasible point can be used. The equivalence means that local and global minimums of the problems coincide. Nonconvex sets with multivalued Euclidean projections are admitted, and the objective function may be lower semicontinuous. The particular case of convex problems is included. The obtained unconstrained or box constrained problem is solved by a version of the branch and bound method combined with local optimization. In principle, any local optimizer can be used within the branch and bound scheme but in numerical experiments sequential quadratic programming method was successfully used. So the proposed exact penalty method does not assume the existence of the objective function outside the allowable area and does not require the selection of the penalty coefficient.
{"title":"The exact projective penalty method for constrained optimization","authors":"","doi":"10.1007/s10898-023-01350-4","DOIUrl":"https://doi.org/10.1007/s10898-023-01350-4","url":null,"abstract":"<h3>Abstract</h3> <p>A new exact projective penalty method is proposed for the equivalent reduction of constrained optimization problems to nonsmooth unconstrained ones. In the method, the original objective function is extended to infeasible points by summing its value at the projection of an infeasible point on the feasible set with the distance to the projection. Beside Euclidean projections, also a pointed projection in the direction of some fixed internal feasible point can be used. The equivalence means that local and global minimums of the problems coincide. Nonconvex sets with multivalued Euclidean projections are admitted, and the objective function may be lower semicontinuous. The particular case of convex problems is included. The obtained unconstrained or box constrained problem is solved by a version of the branch and bound method combined with local optimization. In principle, any local optimizer can be used within the branch and bound scheme but in numerical experiments sequential quadratic programming method was successfully used. So the proposed exact penalty method does not assume the existence of the objective function outside the allowable area and does not require the selection of the penalty coefficient.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"30 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139093752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s10898-023-01336-2
Abstract
While constrained, multiobjective optimization is generally very difficult, there is a special case in which such problems can be solved with a simple, elegant branch-and-bound algorithm. This special case is when the objective and constraint functions are Lipschitz continuous with known Lipschitz constants. Given these Lipschitz constants, one can compute lower bounds on the functions over subregions of the search space. This allows one to iteratively partition the search space into rectangles, deleting those rectangles which—based on the lower bounds—contain points that are all provably infeasible or provably dominated by previously sampled point(s). As the algorithm proceeds, the rectangles that have not been deleted provide a tight covering of the Pareto set in the input space. Unfortunately, for black-box optimization this elegant algorithm cannot be applied, as we would not know the Lipschitz constants. In this paper, we show how one can heuristically extend this branch-and-bound algorithm to the case when the problem functions are black-box using an approach similar to that used in the well-known DIRECT global optimization algorithm. We call the resulting method “simDIRECT.” Initial experience with simDIRECT on test problems suggests that it performs similar to, or better than, multiobjective evolutionary algorithms when solving problems with small numbers of variables (up to 12) and a limited number of runs (up to 600).
{"title":"Constrained multiobjective optimization of expensive black-box functions using a heuristic branch-and-bound approach","authors":"","doi":"10.1007/s10898-023-01336-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01336-2","url":null,"abstract":"<h3>Abstract</h3> <p>While constrained, multiobjective optimization is generally very difficult, there is a special case in which such problems can be solved with a simple, elegant branch-and-bound algorithm. This special case is when the objective and constraint functions are Lipschitz continuous with known Lipschitz constants. Given these Lipschitz constants, one can compute lower bounds on the functions over subregions of the search space. This allows one to iteratively partition the search space into rectangles, deleting those rectangles which—based on the lower bounds—contain points that are all provably infeasible or provably dominated by previously sampled point(s). As the algorithm proceeds, the rectangles that have not been deleted provide a tight covering of the Pareto set in the input space. Unfortunately, for black-box optimization this elegant algorithm cannot be applied, as we would not know the Lipschitz constants. In this paper, we show how one can heuristically extend this branch-and-bound algorithm to the case when the problem functions are black-box using an approach similar to that used in the well-known DIRECT global optimization algorithm. We call the resulting method “simDIRECT.” Initial experience with simDIRECT on test problems suggests that it performs similar to, or better than, multiobjective evolutionary algorithms when solving problems with small numbers of variables (up to 12) and a limited number of runs (up to 600). </p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"1 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139080250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s10898-023-01352-2
Daniela Lera, Maria Chiara Nasso, Mikhail Posypkin, Yaroslav D. Sergeyev
In this paper, the problem of approximating and visualizing the solution set of systems of nonlinear inequalities is considered. It is supposed that left-hand parts of the inequalities can be multiextremal and non-differentiable. Thus, traditional local methods using gradients cannot be applied in these circumstances. Problems of this kind arise in many scientific applications, in particular, in finding working spaces of robots where it is necessary to determine not one but all the solutions of the system of nonlinear inequalities. Global optimization algorithms can be taken as an inspiration for developing methods for solving this problem. In this article, two new methods using two different approximations of Peano–Hilbert space-filling curves actively used in global optimization are proposed. Convergence conditions of the new methods are established. Numerical experiments executed on problems regarding finding the working spaces of several robots show a promising performance of the new algorithms.
{"title":"Determining solution set of nonlinear inequalities using space-filling curves for finding working spaces of planar robots","authors":"Daniela Lera, Maria Chiara Nasso, Mikhail Posypkin, Yaroslav D. Sergeyev","doi":"10.1007/s10898-023-01352-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01352-2","url":null,"abstract":"<p>In this paper, the problem of approximating and visualizing the solution set of systems of nonlinear inequalities is considered. It is supposed that left-hand parts of the inequalities can be multiextremal and non-differentiable. Thus, traditional local methods using gradients cannot be applied in these circumstances. Problems of this kind arise in many scientific applications, in particular, in finding working spaces of robots where it is necessary to determine not one but <i>all</i> the solutions of the system of nonlinear inequalities. Global optimization algorithms can be taken as an inspiration for developing methods for solving this problem. In this article, two new methods using two different approximations of Peano–Hilbert space-filling curves actively used in global optimization are proposed. Convergence conditions of the new methods are established. Numerical experiments executed on problems regarding finding the working spaces of several robots show a promising performance of the new algorithms.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"29 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139077175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s10898-023-01344-2
Abstract
We discuss two key problems related to learning and optimization of neural networks: the computation of the adversarial attack for adversarial robustness and approximate optimization of complex functions. We show that both problems can be cast as instances of DC-programming. We give an explicit decomposition of the corresponding functions as differences of convex functions (DC) and report the results of experiments demonstrating the effectiveness of the DCA algorithm applied to these problems.
摘要 我们讨论了与神经网络学习和优化相关的两个关键问题:计算对抗鲁棒性的对抗攻击和复杂函数的近似优化。我们证明,这两个问题都可以作为 DC 编程的实例。我们给出了相应函数作为凸函数差分 (DC) 的明确分解,并报告了实验结果,证明了 DCA 算法应用于这些问题的有效性。
{"title":"DC-programming for neural network optimizations","authors":"","doi":"10.1007/s10898-023-01344-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01344-2","url":null,"abstract":"<h3>Abstract</h3> <p>We discuss two key problems related to learning and optimization of neural networks: the computation of the adversarial attack for adversarial robustness and approximate optimization of complex functions. We show that both problems can be cast as instances of DC-programming. We give an explicit decomposition of the corresponding functions as differences of convex functions (DC) and report the results of experiments demonstrating the effectiveness of the DCA algorithm applied to these problems. </p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"25 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139080275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-16DOI: 10.1007/s10898-023-01347-z
Quoc Tran-Dinh
We develop two “Nesterov’s accelerated” variants of the well-known extragradient method to approximate a solution of a co-hypomonotone inclusion constituted by the sum of two operators, where one is Lipschitz continuous and the other is possibly multivalued. The first scheme can be viewed as an accelerated variant of Tseng’s forward-backward-forward splitting (FBFS) method, while the second one is a Nesterov’s accelerated variant of the “past” FBFS scheme, which requires only one evaluation of the Lipschitz operator and one resolvent of the multivalued mapping. Under appropriate conditions on the parameters, we theoretically prove that both algorithms achieve (mathcal {O}left( 1/kright) ) last-iterate convergence rates on the residual norm, where k is the iteration counter. Our results can be viewed as alternatives of a recent class of Halpern-type methods for root-finding problems. For comparison, we also provide a new convergence analysis of the two recent extra-anchored gradient-type methods for solving co-hypomonotone inclusions.
我们开发了著名的外梯度法的两个 "涅斯捷罗夫加速 "变体,用于近似求解由两个算子之和构成的共hypomonotone包容体,其中一个算子是立普齐兹连续的,另一个算子可能是多值的。第一种方案可视为曾氏前向-后向-前向分裂(FBFS)方法的加速变体,而第二种方案则是 "过去 "FBFS 方案的涅斯捷罗夫加速变体,只需对 Lipschitz 算子和多值映射的一个解析量进行一次求值。在参数的适当条件下,我们从理论上证明了这两种算法在残差规范上都达到了最后迭代收敛率,其中 k 是迭代计数器。我们的结果可以看作是最近一类用于寻根问题的哈尔彭类方法的替代方案。为了进行比较,我们还对最近的两种用于求解共假单调夹杂的外锚定梯度型方法进行了新的收敛分析。
{"title":"Extragradient-type methods with $$mathcal {O}left( 1/kright) $$ last-iterate convergence rates for co-hypomonotone inclusions","authors":"Quoc Tran-Dinh","doi":"10.1007/s10898-023-01347-z","DOIUrl":"https://doi.org/10.1007/s10898-023-01347-z","url":null,"abstract":"<p>We develop two “Nesterov’s accelerated” variants of the well-known extragradient method to approximate a solution of a co-hypomonotone inclusion constituted by the sum of two operators, where one is Lipschitz continuous and the other is possibly multivalued. The first scheme can be viewed as an accelerated variant of Tseng’s forward-backward-forward splitting (FBFS) method, while the second one is a Nesterov’s accelerated variant of the “past” FBFS scheme, which requires only one evaluation of the Lipschitz operator and one resolvent of the multivalued mapping. Under appropriate conditions on the parameters, we theoretically prove that both algorithms achieve <span>(mathcal {O}left( 1/kright) )</span> last-iterate convergence rates on the residual norm, where <i>k</i> is the iteration counter. Our results can be viewed as alternatives of a recent class of Halpern-type methods for root-finding problems. For comparison, we also provide a new convergence analysis of the two recent extra-anchored gradient-type methods for solving co-hypomonotone inclusions.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"7 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2023-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138681693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-16DOI: 10.1007/s10898-023-01348-y
Xianfu Wang, Ziyuan Wang
We propose a Bregman inertial forward-reflected-backward (BiFRB) method for nonconvex composite problems. Assuming the generalized concave Kurdyka-Łojasiewicz property, we obtain sequential convergence of BiFRB, as well as convergence rates on both the function value and actual sequence. One distinguishing feature in our analysis is that we utilize a careful treatment of merit function parameters, circumventing the usual restrictive assumption on the inertial parameters. We also present formulae for the Bregman subproblem, supplementing not only BiFRB but also the work of Boţ-Csetnek-László and Boţ-Csetnek. Numerical simulations are conducted to evaluate the performance of our proposed algorithm.
{"title":"A Bregman inertial forward-reflected-backward method for nonconvex minimization","authors":"Xianfu Wang, Ziyuan Wang","doi":"10.1007/s10898-023-01348-y","DOIUrl":"https://doi.org/10.1007/s10898-023-01348-y","url":null,"abstract":"<p>We propose a Bregman inertial forward-reflected-backward (BiFRB) method for nonconvex composite problems. Assuming the generalized concave Kurdyka-Łojasiewicz property, we obtain sequential convergence of BiFRB, as well as convergence rates on both the function value and actual sequence. One distinguishing feature in our analysis is that we utilize a careful treatment of merit function parameters, circumventing the usual restrictive assumption on the inertial parameters. We also present formulae for the Bregman subproblem, supplementing not only BiFRB but also the work of Boţ-Csetnek-László and Boţ-Csetnek. Numerical simulations are conducted to evaluate the performance of our proposed algorithm.\u0000</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"8 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2023-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138681761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}