Pub Date : 2024-03-22DOI: 10.1007/s10107-024-02073-4
Alex L. Wang, Fatma Kılınç-Karzan
This paper introduces a new storage-optimal first-order method, CertSDP, for solving a special class of semidefinite programs (SDPs) to high accuracy. The class of SDPs that we consider, the exact QMP-like SDPs, is characterized by low-rank solutions, a priori knowledge of the restriction of the SDP solution to a small subspace, and standard regularity assumptions such as strict complementarity. Crucially, we show how to use a certificate of strict complementarity to construct a low-dimensional strongly convex minimax problem whose optimizer coincides with a factorization of the SDP optimizer. From an algorithmic standpoint, we show how to construct the necessary certificate and how to solve the minimax problem efficiently. Our algorithms for strongly convex minimax problems with inexact prox maps may be of independent interest. We accompany our theoretical results with preliminary numerical experiments suggesting that CertSDP significantly outperforms current state-of-the-art methods on large sparse exact QMP-like SDPs.
{"title":"Accelerated first-order methods for a class of semidefinite programs","authors":"Alex L. Wang, Fatma Kılınç-Karzan","doi":"10.1007/s10107-024-02073-4","DOIUrl":"https://doi.org/10.1007/s10107-024-02073-4","url":null,"abstract":"<p>This paper introduces a new storage-optimal first-order method, CertSDP, for solving a special class of semidefinite programs (SDPs) to high accuracy. The class of SDPs that we consider, the <i>exact QMP-like SDPs</i>, is characterized by low-rank solutions, <i>a priori</i> knowledge of the restriction of the SDP solution to a small subspace, and standard regularity assumptions such as strict complementarity. Crucially, we show how to use a <i>certificate of strict complementarity</i> to construct a low-dimensional strongly convex minimax problem whose optimizer coincides with a factorization of the SDP optimizer. From an algorithmic standpoint, we show how to construct the necessary certificate and how to solve the minimax problem efficiently. Our algorithms for strongly convex minimax problems with inexact prox maps may be of independent interest. We accompany our theoretical results with preliminary numerical experiments suggesting that CertSDP significantly outperforms current state-of-the-art methods on large sparse exact QMP-like SDPs.\u0000</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"31 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140203212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-15DOI: 10.1007/s10107-024-02072-5
Abstract
We consider min–max optimization problems for polynomial functions, where a multivariate polynomial is maximized with respect to a subset of variables, and the resulting maximal value is minimized with respect to the remaining variables. When the variables belong to simple sets (e.g., a hypercube, the Euclidean hypersphere, or a ball), we derive a sum-of-squares formulation based on a primal-dual approach. In the simplest setting, we provide a convergence proof when the degree of the relaxation tends to infinity and observe empirically that it can be finitely convergent in several situations. Moreover, our formulation leads to an interesting link with feasibility certificates for polynomial inequalities based on Putinar’s Positivstellensatz.
{"title":"Sum-of-squares relaxations for polynomial min–max problems over simple sets","authors":"","doi":"10.1007/s10107-024-02072-5","DOIUrl":"https://doi.org/10.1007/s10107-024-02072-5","url":null,"abstract":"<h3>Abstract</h3> <p>We consider min–max optimization problems for polynomial functions, where a multivariate polynomial is maximized with respect to a subset of variables, and the resulting maximal value is minimized with respect to the remaining variables. When the variables belong to simple sets (e.g., a hypercube, the Euclidean hypersphere, or a ball), we derive a sum-of-squares formulation based on a primal-dual approach. In the simplest setting, we provide a convergence proof when the degree of the relaxation tends to infinity and observe empirically that it can be finitely convergent in several situations. Moreover, our formulation leads to an interesting link with feasibility certificates for polynomial inequalities based on Putinar’s Positivstellensatz. </p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"3 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140147007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-14DOI: 10.1007/s10107-024-02076-1
Samuel Burer
Globally optimizing a nonconvex quadratic over the intersection of m balls in (mathbb {R}^n) is known to be polynomial-time solvable for fixed m. Moreover, when (m=1), the standard semidefinite relaxation is exact. When (m=2), it has been shown recently that an exact relaxation can be constructed using a disjunctive semidefinite formulation based essentially on two copies of the (m=1) case. However, there is no known explicit, tractable, exact convex representation for (m ge 3). In this paper, we construct a new, polynomially sized semidefinite relaxation for all m, which does not employ a disjunctive approach. We show that our relaxation is exact for (m=2). Then, for (m ge 3), we demonstrate empirically that it is fast and strong compared to existing relaxations. The key idea of the relaxation is a simple lifting of the original problem into dimension (n, +, 1). Extending this construction: (i) we show that nonconvex quadratic programming over (Vert xVert le min { 1, g + h^T x }) has an exact semidefinite representation; and (ii) we construct a new relaxation for quadratic programming over the intersection of two ellipsoids, which globally solves all instances of a benchmark collection from the literature.
众所周知,对于固定的 m,在 (mathbb {R}^n)中的 m 个球的交点上进行非凸二次函数的全局优化是多项式时间可解的。此外,当 (m=1) 时,标准的半有限松弛是精确的。当(m=2)时,最近有研究表明,可以使用一种基于(m=1)情况的两份分条件半定式来构造精确松弛。然而,对于 (m ge 3) 还没有已知的明确的、可操作的、精确的凸表示。在本文中,我们为所有 m 构建了一个新的、多项式大小的半有限松弛,它没有采用析取方法。我们证明我们的松弛对于(m=2)是精确的。然后,对于 (m ge 3), 我们通过经验证明,与现有的松弛方法相比,我们的松弛方法既快又强。松弛的关键思想是将原始问题简单地提升到维度(n, +, 1)。扩展这个构造:(i)我们证明了在(Vert xVert le min { 1, g + h^T x })上的非凸二次规划有一个精确的半有限表示;(ii)我们为在两个椭圆的交点上的二次规划构造了一个新的松弛,它在全局上解决了文献中一个基准集合的所有实例。
{"title":"A slightly lifted convex relaxation for nonconvex quadratic programming with ball constraints","authors":"Samuel Burer","doi":"10.1007/s10107-024-02076-1","DOIUrl":"https://doi.org/10.1007/s10107-024-02076-1","url":null,"abstract":"<p>Globally optimizing a nonconvex quadratic over the intersection of <i>m</i> balls in <span>(mathbb {R}^n)</span> is known to be polynomial-time solvable for fixed <i>m</i>. Moreover, when <span>(m=1)</span>, the standard semidefinite relaxation is exact. When <span>(m=2)</span>, it has been shown recently that an exact relaxation can be constructed using a disjunctive semidefinite formulation based essentially on two copies of the <span>(m=1)</span> case. However, there is no known explicit, tractable, exact convex representation for <span>(m ge 3)</span>. In this paper, we construct a new, polynomially sized semidefinite relaxation for all <i>m</i>, which does not employ a disjunctive approach. We show that our relaxation is exact for <span>(m=2)</span>. Then, for <span>(m ge 3)</span>, we demonstrate empirically that it is fast and strong compared to existing relaxations. The key idea of the relaxation is a simple lifting of the original problem into dimension <span>(n, +, 1)</span>. Extending this construction: (i) we show that nonconvex quadratic programming over <span>(Vert xVert le min { 1, g + h^T x })</span> has an exact semidefinite representation; and (ii) we construct a new relaxation for quadratic programming over the intersection of two ellipsoids, which globally solves all instances of a benchmark collection from the literature.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"35 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140146947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-13DOI: 10.1007/s10107-024-02075-2
Tianyi Lin, Michael I. Jordan
This paper settles an open and challenging question pertaining to the design of simple and optimal high-order methods for solving smooth and monotone variational inequalities (VIs). A VI involves finding (x^star in {mathcal {X}}) such that (langle F(x), x - x^star rangle ge 0) for all (x in {mathcal {X}}). We consider the setting in which (F: {mathbb {R}}^d rightarrow {mathbb {R}}^d) is smooth with up to ((p-1)^{text {th}})-order derivatives. For (p = 2), the cubic regularization of Newton’s method has been extended to VIs with a global rate of (O(epsilon ^{-1})) (Nesterov in Cubic regularization of Newton’s method for convex problems with constraints, Tech. rep., Université catholique de Louvain, Center for Operations Research and Econometrics (CORE), 2006). An improved rate of (O(epsilon ^{-2/3}log log (1/epsilon ))) can be obtained via an alternative second-order method, but this method requires a nontrivial line-search procedure as an inner loop. Similarly, the existing high-order methods based on line-search procedures have been shown to achieve a rate of (O(epsilon ^{-2/(p+1)}log log (1/epsilon ))) (Bullins and Lai in SIAM J Optim 32(3):2208–2229, 2022; Jiang and Mokhtari in Generalized optimistic methods for convex–concave saddle point problems, 2022; Lin and Jordan in Math Oper Res 48(4):2353–2382, 2023). As emphasized by Nesterov (Lectures on convex optimization, vol 137, Springer, Berlin, 2018), however, such procedures do not necessarily imply the practical applicability in large-scale applications, and it is desirable to complement these results with a simple high-order VI method that retains the optimality of the more complex methods. We propose a (p^{text {th}})-order method that does not require any line search procedure and provably converges to a weak solution at a rate of (O(epsilon ^{-2/(p+1)})). We prove that our (p^{text {th}})-order method is optimal in the monotone setting by establishing a lower bound of (Omega (epsilon ^{-2/(p+1)})) under a generalized linear span assumption. A restarted version of our (p^{text {th}})-order method attains a linear rate for smooth and (p^{text {th}})-order uniformly monotone VIs and another restarted version of our (p^{text {th}})-order method attains a local superlinear rate for smooth and strongly monotone VIs. Further, the similar (p^{text {th}})-order method achieves a global rate of (O(epsilon ^{-2/p})) for solving smooth and nonmonotone VIs satisfying the Minty condition. Two restarted versions attain a global linear rate under additional (p^{text {th}})-order uniform Minty condition and a local superlinear rate under additional strong Minty condition.
本文解决了一个开放且具有挑战性的问题,即设计简单且最优的高阶方法来求解平滑且单调的变分不等式(VIs)。变分不等式涉及找到 (x^starin {mathcal {X}}) such that (angle F(x), x - x^starrangle ge 0) for all (x in {mathcal {X}}).我们考虑这样一种情况:F: {mathbb {R}}^d rightarrow {mathbb {R}}^d) 是光滑的,最多有((p-1)^{text {th}})阶导数。对于 (p = 2), 牛顿方法的立方正则化已经扩展到 VIs,其全局速率为 (O(epsilon ^{-1}))(Nesterov 在 Cubic regularization of Newton's method for convex problems with constraints, Tech. rep.通过另一种二阶方法可以得到一个改进的速率(O(epsilon ^{-2/3}log log (1/epsilon ))) ,但这种方法需要一个非线性的线性搜索过程作为内循环。同样,现有的基于线性搜索过程的高阶方法已经被证明可以达到 (O(epsilon ^{-2/(p+1)}log log (1/epsilon ))) (Bullins 和 Lai 在 SIAM J Optim 32(3):2208-2229, 2022; Jiang and Mokhtari in Generalized optimistic methods for convex-concave saddle point problems, 2022; Lin and Jordan in Math Oper Res 48(4):2353-2382, 2023)。然而,正如涅斯捷罗夫(Lectures on convex optimization, vol 137, Springer, Berlin, 2018)所强调的,这些程序并不一定意味着在大规模应用中的实际适用性,我们希望用一种简单的高阶 VI 方法来补充这些结果,同时保留更复杂方法的最优性。我们提出了一种 (p^{text {th}})阶方法,它不需要任何线性搜索过程,并能以 (O(epsilon ^{-2/(p+1)})) 的速率收敛到弱解。我们通过在广义线性跨度假设下建立一个 (Omega (epsilon ^{-2/(p+1)})) 的下限,证明我们的 (p^{text {th}})-order 方法在单调设置中是最优的。我们的(p^{text {th}})阶方法的重启版本对于平滑和(p^{text {th}})阶均匀单调VI达到了线性速率,我们的(p^{text {th}})阶方法的另一个重启版本对于平滑和强单调VI达到了局部超线性速率。此外,类似的(p^{text {th}})阶方法在求解满足Minty条件的平滑和非单调VI时达到了(O(epsilon ^{-2/p}))的全局速率。两个重启版本在附加的 (p^{text {th}})-order uniform Minty 条件下达到了全局线性速率,在附加的强 Minty 条件下达到了局部超线性速率。
{"title":"Perseus: a simple and optimal high-order method for variational inequalities","authors":"Tianyi Lin, Michael I. Jordan","doi":"10.1007/s10107-024-02075-2","DOIUrl":"https://doi.org/10.1007/s10107-024-02075-2","url":null,"abstract":"<p>This paper settles an open and challenging question pertaining to the design of simple and optimal high-order methods for solving smooth and monotone variational inequalities (VIs). A VI involves finding <span>(x^star in {mathcal {X}})</span> such that <span>(langle F(x), x - x^star rangle ge 0)</span> for all <span>(x in {mathcal {X}})</span>. We consider the setting in which <span>(F: {mathbb {R}}^d rightarrow {mathbb {R}}^d)</span> is smooth with up to <span>((p-1)^{text {th}})</span>-order derivatives. For <span>(p = 2)</span>, the cubic regularization of Newton’s method has been extended to VIs with a global rate of <span>(O(epsilon ^{-1}))</span> (Nesterov in Cubic regularization of Newton’s method for convex problems with constraints, Tech. rep., Université catholique de Louvain, Center for Operations Research and Econometrics (CORE), 2006). An improved rate of <span>(O(epsilon ^{-2/3}log log (1/epsilon )))</span> can be obtained via an alternative second-order method, but this method requires a nontrivial line-search procedure as an inner loop. Similarly, the existing high-order methods based on line-search procedures have been shown to achieve a rate of <span>(O(epsilon ^{-2/(p+1)}log log (1/epsilon )))</span> (Bullins and Lai in SIAM J Optim 32(3):2208–2229, 2022; Jiang and Mokhtari in Generalized optimistic methods for convex–concave saddle point problems, 2022; Lin and Jordan in Math Oper Res 48(4):2353–2382, 2023). As emphasized by Nesterov (Lectures on convex optimization, vol 137, Springer, Berlin, 2018), however, such procedures do not necessarily imply the practical applicability in large-scale applications, and it is desirable to complement these results with a simple high-order VI method that retains the optimality of the more complex methods. We propose a <span>(p^{text {th}})</span>-order method that does <i>not</i> require any line search procedure and provably converges to a weak solution at a rate of <span>(O(epsilon ^{-2/(p+1)}))</span>. We prove that our <span>(p^{text {th}})</span>-order method is optimal in the monotone setting by establishing a lower bound of <span>(Omega (epsilon ^{-2/(p+1)}))</span> under a generalized linear span assumption. A restarted version of our <span>(p^{text {th}})</span>-order method attains a linear rate for smooth and <span>(p^{text {th}})</span>-order uniformly monotone VIs and another restarted version of our <span>(p^{text {th}})</span>-order method attains a local superlinear rate for smooth and strongly monotone VIs. Further, the similar <span>(p^{text {th}})</span>-order method achieves a global rate of <span>(O(epsilon ^{-2/p}))</span> for solving smooth and nonmonotone VIs satisfying the Minty condition. Two restarted versions attain a global linear rate under additional <span>(p^{text {th}})</span>-order uniform Minty condition and a local superlinear rate under additional strong Minty condition.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"16 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140128068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-13DOI: 10.1007/s10107-024-02063-6
Abstract
We consider generalized Nash equilibrium problems (GNEPs) with non-convex strategy spaces and non-convex cost functions. This general class of games includes the important case of games with mixed-integer variables for which only a few results are known in the literature. We present a new approach to characterize equilibria via a convexification technique using the Nikaido–Isoda function. To any given instance of the GNEP, we construct a set of convexified instances and show that a feasible strategy profile is an equilibrium for the original instance if and only if it is an equilibrium for any convexified instance and the convexified cost functions coincide with the initial ones. We develop this convexification approach along three dimensions: We first show that for quasi-linear models, where a convexified instance exists in which for fixed strategies of the opponent players, the cost function of every player is linear and the respective strategy space is polyhedral, the convexification reduces the GNEP to a standard (non-linear) optimization problem. Secondly, we derive two complete characterizations of those GNEPs for which the convexification leads to a jointly constrained or a jointly convex GNEP, respectively. These characterizations require new concepts related to the interplay of the convex hull operator applied to restricted subsets of feasible strategies and may be interesting on their own. Note that this characterization is also computationally relevant as jointly convex GNEPs have been extensively studied in the literature. Finally, we demonstrate the applicability of our results by presenting a numerical study regarding the computation of equilibria for three classes of GNEPs related to integral network flows and discrete market equilibria.
{"title":"Generalized Nash equilibrium problems with mixed-integer variables","authors":"","doi":"10.1007/s10107-024-02063-6","DOIUrl":"https://doi.org/10.1007/s10107-024-02063-6","url":null,"abstract":"<h3>Abstract</h3> <p>We consider generalized Nash equilibrium problems (GNEPs) with non-convex strategy spaces and non-convex cost functions. This general class of games includes the important case of games with mixed-integer variables for which only a few results are known in the literature. We present a new approach to characterize equilibria via a convexification technique using the Nikaido–Isoda function. To any given instance of the GNEP, we construct a set of convexified instances and show that a feasible strategy profile is an equilibrium for the original instance if and only if it is an equilibrium for any convexified instance and the convexified cost functions coincide with the initial ones. We develop this convexification approach along three dimensions: We first show that for <em>quasi-linear</em> models, where a convexified instance exists in which for fixed strategies of the opponent players, the cost function of every player is linear and the respective strategy space is polyhedral, the convexification reduces the GNEP to a standard (non-linear) optimization problem. Secondly, we derive two complete characterizations of those GNEPs for which the convexification leads to a jointly constrained or a jointly convex GNEP, respectively. These characterizations require new concepts related to the interplay of the convex hull operator applied to restricted subsets of feasible strategies and may be interesting on their own. Note that this characterization is also computationally relevant as jointly convex GNEPs have been extensively studied in the literature. Finally, we demonstrate the applicability of our results by presenting a numerical study regarding the computation of equilibria for three classes of GNEPs related to integral network flows and discrete market equilibria.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"109 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140128067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-13DOI: 10.1007/s10107-024-02069-0
Abstract
In this paper we study the well-known Chvátal–Gomory (CG) procedure for the class of integer semidefinite programs (ISDPs). We prove several results regarding the hierarchy of relaxations obtained by iterating this procedure. We also study different formulations of the elementary closure of spectrahedra. A polyhedral description of the elementary closure for a specific type of spectrahedra is derived by exploiting total dual integrality for SDPs. Moreover, we show how to exploit (strengthened) CG cuts in a branch-and-cut framework for ISDPs. Different from existing algorithms in the literature, the separation routine in our approach exploits both the semidefinite and the integrality constraints. We provide separation routines for several common classes of binary SDPs resulting from combinatorial optimization problems. In the second part of the paper we present a comprehensive application of our approach to the quadratic traveling salesman problem (QTSP). Based on the algebraic connectivity of the directed Hamiltonian cycle, two ISDPs that model the QTSP are introduced. We show that the CG cuts resulting from these formulations contain several well-known families of cutting planes. Numerical results illustrate the practical strength of the CG cuts in our branch-and-cut algorithm, which outperforms alternative ISDP solvers and is able to solve large QTSP instances to optimality.
{"title":"The Chvátal–Gomory procedure for integer SDPs with applications in combinatorial optimization","authors":"","doi":"10.1007/s10107-024-02069-0","DOIUrl":"https://doi.org/10.1007/s10107-024-02069-0","url":null,"abstract":"<h3>Abstract</h3> <p>In this paper we study the well-known Chvátal–Gomory (CG) procedure for the class of integer semidefinite programs (ISDPs). We prove several results regarding the hierarchy of relaxations obtained by iterating this procedure. We also study different formulations of the elementary closure of spectrahedra. A polyhedral description of the elementary closure for a specific type of spectrahedra is derived by exploiting total dual integrality for SDPs. Moreover, we show how to exploit (strengthened) CG cuts in a branch-and-cut framework for ISDPs. Different from existing algorithms in the literature, the separation routine in our approach exploits both the semidefinite and the integrality constraints. We provide separation routines for several common classes of binary SDPs resulting from combinatorial optimization problems. In the second part of the paper we present a comprehensive application of our approach to the quadratic traveling salesman problem (<span>QTSP</span>). Based on the algebraic connectivity of the directed Hamiltonian cycle, two ISDPs that model the <span>QTSP</span> are introduced. We show that the CG cuts resulting from these formulations contain several well-known families of cutting planes. Numerical results illustrate the practical strength of the CG cuts in our branch-and-cut algorithm, which outperforms alternative ISDP solvers and is able to solve large <span>QTSP</span> instances to optimality.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"2017 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140127998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-07DOI: 10.1007/s10107-024-02064-5
Martin Dvorak, Vladimir Kolmogorov
Given a fixed finite metric space ((V,mu )), the minimum 0-extension problem, denoted as (mathtt{0hbox {-}Ext}[{mu }]), is equivalent to the following optimization problem: minimize function of the form (min nolimits _{xin V^n} sum _i f_i(x_i) + sum _{ij} c_{ij}hspace{0.5pt}mu (x_i,x_j)) where (f_i:Vrightarrow mathbb {R}) are functions given by (f_i(x_i)=sum _{vin V} c_{vi}hspace{0.5pt}mu (x_i,v)) and (c_{ij},c_{vi}) are given nonnegative costs. The computational complexity of (mathtt{0hbox {-}Ext}[{mu }]) has been recently established by Karzanov and by Hirai: if metric (mu ) is orientable modular then (mathtt{0hbox {-}Ext}[{mu }]) can be solved in polynomial time, otherwise (mathtt{0hbox {-}Ext}[{mu }]) is NP-hard. To prove the tractability part, Hirai developed a theory of discrete convex functions on orientable modular graphs generalizing several known classes of functions in discrete convex analysis, such as (L^natural )-convex functions. We consider a more general version of the problem in which unary functions (f_i(x_i)) can additionally have terms of the form (c_{uv;i}hspace{0.5pt}mu (x_i,{u,v})) for ({u,!hspace{0.5pt}hspace{0.5pt}v}in F), where set (Fsubseteq left( {begin{array}{c}V 2end{array}}right) ) is fixed. We extend the complexity classification above by providing an explicit condition on ((mu ,F)) for the problem to be tractable. In order to prove the tractability part, we generalize Hirai’s theory and define a larger class of discrete convex functions. It covers, in particular, another well-known class of functions, namely submodular functions on an integer lattice. Finally, we improve the complexity of Hirai’s algorithm for solving (mathtt{0hbox {-}Ext}[{mu }]) on orientable modular graphs.
{"title":"Generalized minimum 0-extension problem and discrete convexity","authors":"Martin Dvorak, Vladimir Kolmogorov","doi":"10.1007/s10107-024-02064-5","DOIUrl":"https://doi.org/10.1007/s10107-024-02064-5","url":null,"abstract":"<p>Given a fixed finite metric space <span>((V,mu ))</span>, the <i>minimum 0-extension problem</i>, denoted as <span>(mathtt{0hbox {-}Ext}[{mu }])</span>, is equivalent to the following optimization problem: minimize function of the form <span>(min nolimits _{xin V^n} sum _i f_i(x_i) + sum _{ij} c_{ij}hspace{0.5pt}mu (x_i,x_j))</span> where <span>(f_i:Vrightarrow mathbb {R})</span> are functions given by <span>(f_i(x_i)=sum _{vin V} c_{vi}hspace{0.5pt}mu (x_i,v))</span> and <span>(c_{ij},c_{vi})</span> are given nonnegative costs. The computational complexity of <span>(mathtt{0hbox {-}Ext}[{mu }])</span> has been recently established by Karzanov and by Hirai: if metric <span>(mu )</span> is <i>orientable modular</i> then <span>(mathtt{0hbox {-}Ext}[{mu }])</span> can be solved in polynomial time, otherwise <span>(mathtt{0hbox {-}Ext}[{mu }])</span> is NP-hard. To prove the tractability part, Hirai developed a theory of discrete convex functions on orientable modular graphs generalizing several known classes of functions in discrete convex analysis, such as <span>(L^natural )</span>-convex functions. We consider a more general version of the problem in which unary functions <span>(f_i(x_i))</span> can additionally have terms of the form <span>(c_{uv;i}hspace{0.5pt}mu (x_i,{u,v}))</span> for <span>({u,!hspace{0.5pt}hspace{0.5pt}v}in F)</span>, where set <span>(Fsubseteq left( {begin{array}{c}V 2end{array}}right) )</span> is fixed. We extend the complexity classification above by providing an explicit condition on <span>((mu ,F))</span> for the problem to be tractable. In order to prove the tractability part, we generalize Hirai’s theory and define a larger class of discrete convex functions. It covers, in particular, another well-known class of functions, namely submodular functions on an integer lattice. Finally, we improve the complexity of Hirai’s algorithm for solving <span>(mathtt{0hbox {-}Ext}[{mu }])</span> on orientable modular graphs.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"87 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140074954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02057-4
Digvijay Boob, Qi Deng, Guanghui Lan
We present a new feasible proximal gradient method for constrained optimization where both the objective and constraint functions are given by summation of a smooth, possibly nonconvex function and a convex simple function. The algorithm converts the original problem into a sequence of convex subproblems. Formulating those subproblems requires the evaluation of at most one gradient-value of the original objective and constraint functions. Either exact or approximate subproblems solutions can be computed efficiently in many cases. An important feature of the algorithm is the constraint level parameter. By carefully increasing this level for each subproblem, we provide a simple solution to overcome the challenge of bounding the Lagrangian multipliers and show that the algorithm follows a strictly feasible solution path till convergence to the stationary point. We develop a simple, proximal gradient descent type analysis, showing that the complexity bound of this new algorithm is comparable to gradient descent for the unconstrained setting which is new in the literature. Exploiting this new design and analysis technique, we extend our algorithms to some more challenging constrained optimization problems where (1) the objective is a stochastic or finite-sum function, and (2) structured nonsmooth functions replace smooth components of both objective and constraint functions. Complexity results for these problems also seem to be new in the literature. Finally, our method can also be applied to convex function constrained problems where we show complexities similar to the proximal gradient method.
{"title":"Level constrained first order methods for function constrained optimization","authors":"Digvijay Boob, Qi Deng, Guanghui Lan","doi":"10.1007/s10107-024-02057-4","DOIUrl":"https://doi.org/10.1007/s10107-024-02057-4","url":null,"abstract":"<p>We present a new feasible proximal gradient method for constrained optimization where both the objective and constraint functions are given by summation of a smooth, possibly nonconvex function and a convex simple function. The algorithm converts the original problem into a sequence of convex subproblems. Formulating those subproblems requires the evaluation of at most one gradient-value of the original objective and constraint functions. Either exact or approximate subproblems solutions can be computed efficiently in many cases. An important feature of the algorithm is the constraint level parameter. By carefully increasing this level for each subproblem, we provide a simple solution to overcome the challenge of bounding the Lagrangian multipliers and show that the algorithm follows a strictly feasible solution path till convergence to the stationary point. We develop a simple, proximal gradient descent type analysis, showing that the complexity bound of this new algorithm is comparable to gradient descent for the unconstrained setting which is new in the literature. Exploiting this new design and analysis technique, we extend our algorithms to some more challenging constrained optimization problems where (1) the objective is a stochastic or finite-sum function, and (2) structured nonsmooth functions replace smooth components of both objective and constraint functions. Complexity results for these problems also seem to be new in the literature. Finally, our method can also be applied to convex function constrained problems where we show complexities similar to the proximal gradient method.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"43 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140054873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02070-7
Yuzhou Qiu, E. Alper Yıldırım
We study linear programming relaxations of nonconvex quadratic programs given by the reformulation–linearization technique (RLT), referred to as RLT relaxations. We investigate the relations between the polyhedral properties of the feasible regions of a quadratic program and its RLT relaxation. We establish various connections between recession directions, boundedness, and vertices of the two feasible regions. Using these properties, we present a complete description of the set of instances that admit an exact RLT relaxation. We then give a thorough discussion of how our results can be converted into simple algorithmic procedures to construct instances of quadratic programs with exact, inexact, or unbounded RLT relaxations.
{"title":"Polyhedral properties of RLT relaxations of nonconvex quadratic programs and their implications on exact relaxations","authors":"Yuzhou Qiu, E. Alper Yıldırım","doi":"10.1007/s10107-024-02070-7","DOIUrl":"https://doi.org/10.1007/s10107-024-02070-7","url":null,"abstract":"<p>We study linear programming relaxations of nonconvex quadratic programs given by the reformulation–linearization technique (RLT), referred to as RLT relaxations. We investigate the relations between the polyhedral properties of the feasible regions of a quadratic program and its RLT relaxation. We establish various connections between recession directions, boundedness, and vertices of the two feasible regions. Using these properties, we present a complete description of the set of instances that admit an exact RLT relaxation. We then give a thorough discussion of how our results can be converted into simple algorithmic procedures to construct instances of quadratic programs with exact, inexact, or unbounded RLT relaxations.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"11 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140057690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02068-1
Yaohua Hu, Xinlin Hu, Xiaoqi Yang
This paper aims to find an approximate true sparse solution of an underdetermined linear system. For this purpose, we propose two types of iterative thresholding algorithms with the continuation technique and the truncation technique respectively. We introduce a notion of limited shrinkage thresholding operator and apply it, together with the restricted isometry property, to show that the proposed algorithms converge to an approximate true sparse solution within a tolerance relevant to the noise level and the limited shrinkage magnitude. Applying the obtained results to nonconvex regularization problems with SCAD, MCP and (ell _p) penalty ((0le p le 1)) and utilizing the recovery bound theory, we establish the convergence of their proximal gradient algorithms to an approximate global solution of nonconvex regularization problems. The established results include the existing convergence theory for (ell _1) or (ell _0) regularization problems for finding a true sparse solution as special cases. Preliminary numerical results show that our proposed algorithms can find approximate true sparse solutions that are much better than stationary solutions that are found by using the standard proximal gradient algorithm.
本文旨在寻找未定线性系统的近似真稀疏解。为此,我们提出了两种迭代阈值算法,分别采用延续技术和截断技术。我们引入了有限收缩阈值算子的概念,并将其与受限等距特性一起应用,证明所提出的算法能在与噪声水平和有限收缩幅度相关的容差范围内收敛到近似真稀疏解。将得到的结果应用于具有 SCAD、MCP 和 (ell _p)惩罚((0le p le 1))的非凸正则化问题,并利用恢复约束理论,我们建立了它们的近似梯度算法对非凸正则化问题近似全局解的收敛性。所建立的结果包括现有的收敛理论,用于寻找真正稀疏解的(ell _1)或(ell _0)正则化问题。初步的数值结果表明,我们提出的算法可以找到近似的真稀疏解,比使用标准近似梯度算法找到的静态解要好得多。
{"title":"On convergence of iterative thresholding algorithms to approximate sparse solution for composite nonconvex optimization","authors":"Yaohua Hu, Xinlin Hu, Xiaoqi Yang","doi":"10.1007/s10107-024-02068-1","DOIUrl":"https://doi.org/10.1007/s10107-024-02068-1","url":null,"abstract":"<p>This paper aims to find an approximate true sparse solution of an underdetermined linear system. For this purpose, we propose two types of iterative thresholding algorithms with the continuation technique and the truncation technique respectively. We introduce a notion of limited shrinkage thresholding operator and apply it, together with the restricted isometry property, to show that the proposed algorithms converge to an approximate true sparse solution within a tolerance relevant to the noise level and the limited shrinkage magnitude. Applying the obtained results to nonconvex regularization problems with SCAD, MCP and <span>(ell _p)</span> penalty (<span>(0le p le 1)</span>) and utilizing the recovery bound theory, we establish the convergence of their proximal gradient algorithms to an approximate global solution of nonconvex regularization problems. The established results include the existing convergence theory for <span>(ell _1)</span> or <span>(ell _0)</span> regularization problems for finding a true sparse solution as special cases. Preliminary numerical results show that our proposed algorithms can find approximate true sparse solutions that are much better than stationary solutions that are found by using the standard proximal gradient algorithm.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"105 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140054988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}