Pub Date : 2024-03-13DOI: 10.1007/s10107-024-02063-6
Abstract
We consider generalized Nash equilibrium problems (GNEPs) with non-convex strategy spaces and non-convex cost functions. This general class of games includes the important case of games with mixed-integer variables for which only a few results are known in the literature. We present a new approach to characterize equilibria via a convexification technique using the Nikaido–Isoda function. To any given instance of the GNEP, we construct a set of convexified instances and show that a feasible strategy profile is an equilibrium for the original instance if and only if it is an equilibrium for any convexified instance and the convexified cost functions coincide with the initial ones. We develop this convexification approach along three dimensions: We first show that for quasi-linear models, where a convexified instance exists in which for fixed strategies of the opponent players, the cost function of every player is linear and the respective strategy space is polyhedral, the convexification reduces the GNEP to a standard (non-linear) optimization problem. Secondly, we derive two complete characterizations of those GNEPs for which the convexification leads to a jointly constrained or a jointly convex GNEP, respectively. These characterizations require new concepts related to the interplay of the convex hull operator applied to restricted subsets of feasible strategies and may be interesting on their own. Note that this characterization is also computationally relevant as jointly convex GNEPs have been extensively studied in the literature. Finally, we demonstrate the applicability of our results by presenting a numerical study regarding the computation of equilibria for three classes of GNEPs related to integral network flows and discrete market equilibria.
{"title":"Generalized Nash equilibrium problems with mixed-integer variables","authors":"","doi":"10.1007/s10107-024-02063-6","DOIUrl":"https://doi.org/10.1007/s10107-024-02063-6","url":null,"abstract":"<h3>Abstract</h3> <p>We consider generalized Nash equilibrium problems (GNEPs) with non-convex strategy spaces and non-convex cost functions. This general class of games includes the important case of games with mixed-integer variables for which only a few results are known in the literature. We present a new approach to characterize equilibria via a convexification technique using the Nikaido–Isoda function. To any given instance of the GNEP, we construct a set of convexified instances and show that a feasible strategy profile is an equilibrium for the original instance if and only if it is an equilibrium for any convexified instance and the convexified cost functions coincide with the initial ones. We develop this convexification approach along three dimensions: We first show that for <em>quasi-linear</em> models, where a convexified instance exists in which for fixed strategies of the opponent players, the cost function of every player is linear and the respective strategy space is polyhedral, the convexification reduces the GNEP to a standard (non-linear) optimization problem. Secondly, we derive two complete characterizations of those GNEPs for which the convexification leads to a jointly constrained or a jointly convex GNEP, respectively. These characterizations require new concepts related to the interplay of the convex hull operator applied to restricted subsets of feasible strategies and may be interesting on their own. Note that this characterization is also computationally relevant as jointly convex GNEPs have been extensively studied in the literature. Finally, we demonstrate the applicability of our results by presenting a numerical study regarding the computation of equilibria for three classes of GNEPs related to integral network flows and discrete market equilibria.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"109 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140128067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-13DOI: 10.1007/s10107-024-02069-0
Abstract
In this paper we study the well-known Chvátal–Gomory (CG) procedure for the class of integer semidefinite programs (ISDPs). We prove several results regarding the hierarchy of relaxations obtained by iterating this procedure. We also study different formulations of the elementary closure of spectrahedra. A polyhedral description of the elementary closure for a specific type of spectrahedra is derived by exploiting total dual integrality for SDPs. Moreover, we show how to exploit (strengthened) CG cuts in a branch-and-cut framework for ISDPs. Different from existing algorithms in the literature, the separation routine in our approach exploits both the semidefinite and the integrality constraints. We provide separation routines for several common classes of binary SDPs resulting from combinatorial optimization problems. In the second part of the paper we present a comprehensive application of our approach to the quadratic traveling salesman problem (QTSP). Based on the algebraic connectivity of the directed Hamiltonian cycle, two ISDPs that model the QTSP are introduced. We show that the CG cuts resulting from these formulations contain several well-known families of cutting planes. Numerical results illustrate the practical strength of the CG cuts in our branch-and-cut algorithm, which outperforms alternative ISDP solvers and is able to solve large QTSP instances to optimality.
{"title":"The Chvátal–Gomory procedure for integer SDPs with applications in combinatorial optimization","authors":"","doi":"10.1007/s10107-024-02069-0","DOIUrl":"https://doi.org/10.1007/s10107-024-02069-0","url":null,"abstract":"<h3>Abstract</h3> <p>In this paper we study the well-known Chvátal–Gomory (CG) procedure for the class of integer semidefinite programs (ISDPs). We prove several results regarding the hierarchy of relaxations obtained by iterating this procedure. We also study different formulations of the elementary closure of spectrahedra. A polyhedral description of the elementary closure for a specific type of spectrahedra is derived by exploiting total dual integrality for SDPs. Moreover, we show how to exploit (strengthened) CG cuts in a branch-and-cut framework for ISDPs. Different from existing algorithms in the literature, the separation routine in our approach exploits both the semidefinite and the integrality constraints. We provide separation routines for several common classes of binary SDPs resulting from combinatorial optimization problems. In the second part of the paper we present a comprehensive application of our approach to the quadratic traveling salesman problem (<span>QTSP</span>). Based on the algebraic connectivity of the directed Hamiltonian cycle, two ISDPs that model the <span>QTSP</span> are introduced. We show that the CG cuts resulting from these formulations contain several well-known families of cutting planes. Numerical results illustrate the practical strength of the CG cuts in our branch-and-cut algorithm, which outperforms alternative ISDP solvers and is able to solve large <span>QTSP</span> instances to optimality.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"2017 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140127998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-07DOI: 10.1007/s10107-024-02064-5
Martin Dvorak, Vladimir Kolmogorov
Given a fixed finite metric space ((V,mu )), the minimum 0-extension problem, denoted as (mathtt{0hbox {-}Ext}[{mu }]), is equivalent to the following optimization problem: minimize function of the form (min nolimits _{xin V^n} sum _i f_i(x_i) + sum _{ij} c_{ij}hspace{0.5pt}mu (x_i,x_j)) where (f_i:Vrightarrow mathbb {R}) are functions given by (f_i(x_i)=sum _{vin V} c_{vi}hspace{0.5pt}mu (x_i,v)) and (c_{ij},c_{vi}) are given nonnegative costs. The computational complexity of (mathtt{0hbox {-}Ext}[{mu }]) has been recently established by Karzanov and by Hirai: if metric (mu ) is orientable modular then (mathtt{0hbox {-}Ext}[{mu }]) can be solved in polynomial time, otherwise (mathtt{0hbox {-}Ext}[{mu }]) is NP-hard. To prove the tractability part, Hirai developed a theory of discrete convex functions on orientable modular graphs generalizing several known classes of functions in discrete convex analysis, such as (L^natural )-convex functions. We consider a more general version of the problem in which unary functions (f_i(x_i)) can additionally have terms of the form (c_{uv;i}hspace{0.5pt}mu (x_i,{u,v})) for ({u,!hspace{0.5pt}hspace{0.5pt}v}in F), where set (Fsubseteq left( {begin{array}{c}V 2end{array}}right) ) is fixed. We extend the complexity classification above by providing an explicit condition on ((mu ,F)) for the problem to be tractable. In order to prove the tractability part, we generalize Hirai’s theory and define a larger class of discrete convex functions. It covers, in particular, another well-known class of functions, namely submodular functions on an integer lattice. Finally, we improve the complexity of Hirai’s algorithm for solving (mathtt{0hbox {-}Ext}[{mu }]) on orientable modular graphs.
{"title":"Generalized minimum 0-extension problem and discrete convexity","authors":"Martin Dvorak, Vladimir Kolmogorov","doi":"10.1007/s10107-024-02064-5","DOIUrl":"https://doi.org/10.1007/s10107-024-02064-5","url":null,"abstract":"<p>Given a fixed finite metric space <span>((V,mu ))</span>, the <i>minimum 0-extension problem</i>, denoted as <span>(mathtt{0hbox {-}Ext}[{mu }])</span>, is equivalent to the following optimization problem: minimize function of the form <span>(min nolimits _{xin V^n} sum _i f_i(x_i) + sum _{ij} c_{ij}hspace{0.5pt}mu (x_i,x_j))</span> where <span>(f_i:Vrightarrow mathbb {R})</span> are functions given by <span>(f_i(x_i)=sum _{vin V} c_{vi}hspace{0.5pt}mu (x_i,v))</span> and <span>(c_{ij},c_{vi})</span> are given nonnegative costs. The computational complexity of <span>(mathtt{0hbox {-}Ext}[{mu }])</span> has been recently established by Karzanov and by Hirai: if metric <span>(mu )</span> is <i>orientable modular</i> then <span>(mathtt{0hbox {-}Ext}[{mu }])</span> can be solved in polynomial time, otherwise <span>(mathtt{0hbox {-}Ext}[{mu }])</span> is NP-hard. To prove the tractability part, Hirai developed a theory of discrete convex functions on orientable modular graphs generalizing several known classes of functions in discrete convex analysis, such as <span>(L^natural )</span>-convex functions. We consider a more general version of the problem in which unary functions <span>(f_i(x_i))</span> can additionally have terms of the form <span>(c_{uv;i}hspace{0.5pt}mu (x_i,{u,v}))</span> for <span>({u,!hspace{0.5pt}hspace{0.5pt}v}in F)</span>, where set <span>(Fsubseteq left( {begin{array}{c}V 2end{array}}right) )</span> is fixed. We extend the complexity classification above by providing an explicit condition on <span>((mu ,F))</span> for the problem to be tractable. In order to prove the tractability part, we generalize Hirai’s theory and define a larger class of discrete convex functions. It covers, in particular, another well-known class of functions, namely submodular functions on an integer lattice. Finally, we improve the complexity of Hirai’s algorithm for solving <span>(mathtt{0hbox {-}Ext}[{mu }])</span> on orientable modular graphs.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"87 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140074954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02057-4
Digvijay Boob, Qi Deng, Guanghui Lan
We present a new feasible proximal gradient method for constrained optimization where both the objective and constraint functions are given by summation of a smooth, possibly nonconvex function and a convex simple function. The algorithm converts the original problem into a sequence of convex subproblems. Formulating those subproblems requires the evaluation of at most one gradient-value of the original objective and constraint functions. Either exact or approximate subproblems solutions can be computed efficiently in many cases. An important feature of the algorithm is the constraint level parameter. By carefully increasing this level for each subproblem, we provide a simple solution to overcome the challenge of bounding the Lagrangian multipliers and show that the algorithm follows a strictly feasible solution path till convergence to the stationary point. We develop a simple, proximal gradient descent type analysis, showing that the complexity bound of this new algorithm is comparable to gradient descent for the unconstrained setting which is new in the literature. Exploiting this new design and analysis technique, we extend our algorithms to some more challenging constrained optimization problems where (1) the objective is a stochastic or finite-sum function, and (2) structured nonsmooth functions replace smooth components of both objective and constraint functions. Complexity results for these problems also seem to be new in the literature. Finally, our method can also be applied to convex function constrained problems where we show complexities similar to the proximal gradient method.
{"title":"Level constrained first order methods for function constrained optimization","authors":"Digvijay Boob, Qi Deng, Guanghui Lan","doi":"10.1007/s10107-024-02057-4","DOIUrl":"https://doi.org/10.1007/s10107-024-02057-4","url":null,"abstract":"<p>We present a new feasible proximal gradient method for constrained optimization where both the objective and constraint functions are given by summation of a smooth, possibly nonconvex function and a convex simple function. The algorithm converts the original problem into a sequence of convex subproblems. Formulating those subproblems requires the evaluation of at most one gradient-value of the original objective and constraint functions. Either exact or approximate subproblems solutions can be computed efficiently in many cases. An important feature of the algorithm is the constraint level parameter. By carefully increasing this level for each subproblem, we provide a simple solution to overcome the challenge of bounding the Lagrangian multipliers and show that the algorithm follows a strictly feasible solution path till convergence to the stationary point. We develop a simple, proximal gradient descent type analysis, showing that the complexity bound of this new algorithm is comparable to gradient descent for the unconstrained setting which is new in the literature. Exploiting this new design and analysis technique, we extend our algorithms to some more challenging constrained optimization problems where (1) the objective is a stochastic or finite-sum function, and (2) structured nonsmooth functions replace smooth components of both objective and constraint functions. Complexity results for these problems also seem to be new in the literature. Finally, our method can also be applied to convex function constrained problems where we show complexities similar to the proximal gradient method.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"43 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140054873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02070-7
Yuzhou Qiu, E. Alper Yıldırım
We study linear programming relaxations of nonconvex quadratic programs given by the reformulation–linearization technique (RLT), referred to as RLT relaxations. We investigate the relations between the polyhedral properties of the feasible regions of a quadratic program and its RLT relaxation. We establish various connections between recession directions, boundedness, and vertices of the two feasible regions. Using these properties, we present a complete description of the set of instances that admit an exact RLT relaxation. We then give a thorough discussion of how our results can be converted into simple algorithmic procedures to construct instances of quadratic programs with exact, inexact, or unbounded RLT relaxations.
{"title":"Polyhedral properties of RLT relaxations of nonconvex quadratic programs and their implications on exact relaxations","authors":"Yuzhou Qiu, E. Alper Yıldırım","doi":"10.1007/s10107-024-02070-7","DOIUrl":"https://doi.org/10.1007/s10107-024-02070-7","url":null,"abstract":"<p>We study linear programming relaxations of nonconvex quadratic programs given by the reformulation–linearization technique (RLT), referred to as RLT relaxations. We investigate the relations between the polyhedral properties of the feasible regions of a quadratic program and its RLT relaxation. We establish various connections between recession directions, boundedness, and vertices of the two feasible regions. Using these properties, we present a complete description of the set of instances that admit an exact RLT relaxation. We then give a thorough discussion of how our results can be converted into simple algorithmic procedures to construct instances of quadratic programs with exact, inexact, or unbounded RLT relaxations.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"11 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140057690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02068-1
Yaohua Hu, Xinlin Hu, Xiaoqi Yang
This paper aims to find an approximate true sparse solution of an underdetermined linear system. For this purpose, we propose two types of iterative thresholding algorithms with the continuation technique and the truncation technique respectively. We introduce a notion of limited shrinkage thresholding operator and apply it, together with the restricted isometry property, to show that the proposed algorithms converge to an approximate true sparse solution within a tolerance relevant to the noise level and the limited shrinkage magnitude. Applying the obtained results to nonconvex regularization problems with SCAD, MCP and (ell _p) penalty ((0le p le 1)) and utilizing the recovery bound theory, we establish the convergence of their proximal gradient algorithms to an approximate global solution of nonconvex regularization problems. The established results include the existing convergence theory for (ell _1) or (ell _0) regularization problems for finding a true sparse solution as special cases. Preliminary numerical results show that our proposed algorithms can find approximate true sparse solutions that are much better than stationary solutions that are found by using the standard proximal gradient algorithm.
本文旨在寻找未定线性系统的近似真稀疏解。为此,我们提出了两种迭代阈值算法,分别采用延续技术和截断技术。我们引入了有限收缩阈值算子的概念,并将其与受限等距特性一起应用,证明所提出的算法能在与噪声水平和有限收缩幅度相关的容差范围内收敛到近似真稀疏解。将得到的结果应用于具有 SCAD、MCP 和 (ell _p)惩罚((0le p le 1))的非凸正则化问题,并利用恢复约束理论,我们建立了它们的近似梯度算法对非凸正则化问题近似全局解的收敛性。所建立的结果包括现有的收敛理论,用于寻找真正稀疏解的(ell _1)或(ell _0)正则化问题。初步的数值结果表明,我们提出的算法可以找到近似的真稀疏解,比使用标准近似梯度算法找到的静态解要好得多。
{"title":"On convergence of iterative thresholding algorithms to approximate sparse solution for composite nonconvex optimization","authors":"Yaohua Hu, Xinlin Hu, Xiaoqi Yang","doi":"10.1007/s10107-024-02068-1","DOIUrl":"https://doi.org/10.1007/s10107-024-02068-1","url":null,"abstract":"<p>This paper aims to find an approximate true sparse solution of an underdetermined linear system. For this purpose, we propose two types of iterative thresholding algorithms with the continuation technique and the truncation technique respectively. We introduce a notion of limited shrinkage thresholding operator and apply it, together with the restricted isometry property, to show that the proposed algorithms converge to an approximate true sparse solution within a tolerance relevant to the noise level and the limited shrinkage magnitude. Applying the obtained results to nonconvex regularization problems with SCAD, MCP and <span>(ell _p)</span> penalty (<span>(0le p le 1)</span>) and utilizing the recovery bound theory, we establish the convergence of their proximal gradient algorithms to an approximate global solution of nonconvex regularization problems. The established results include the existing convergence theory for <span>(ell _1)</span> or <span>(ell _0)</span> regularization problems for finding a true sparse solution as special cases. Preliminary numerical results show that our proposed algorithms can find approximate true sparse solutions that are much better than stationary solutions that are found by using the standard proximal gradient algorithm.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"105 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140054988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02065-4
Abstract
We show how to round any half-integral solution to the subtour-elimination relaxation for the TSP, while losing a less-than(-) 1.5 factor. Such a rounding algorithm was recently given by Karlin, Klein, and Oveis Gharan based on sampling from max-entropy distributions. We build on an approach of Haddadan and Newman to show how sampling from the matroid intersection polytope, combined with a novel use of max-entropy sampling, can give better guarantees.
{"title":"Matroid-based TSP rounding for half-integral solutions","authors":"","doi":"10.1007/s10107-024-02065-4","DOIUrl":"https://doi.org/10.1007/s10107-024-02065-4","url":null,"abstract":"<h3>Abstract</h3> <p>We show how to round any half-integral solution to the subtour-elimination relaxation for the TSP, while losing a less-than<span> <span>(-)</span> </span> 1.5 factor. Such a rounding algorithm was recently given by Karlin, Klein, and Oveis Gharan based on sampling from max-entropy distributions. We build on an approach of Haddadan and Newman to show how sampling from the matroid intersection polytope, combined with a novel use of max-entropy sampling, can give better guarantees.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"274 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140057732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10107-024-02066-3
Abstract
We present an auction algorithm using multiplicative instead of constant weight updates to compute a ((1-varepsilon ))-approximate maximum weight matching (MWM) in a bipartite graph with n vertices and m edges in time (O(mvarepsilon ^{-1})), beating the running time of the fastest known approximation algorithm of Duan and Pettie [JACM ’14] that runs in (O(mvarepsilon ^{-1}log varepsilon ^{-1})). Our algorithm is very simple and it can be extended to give a dynamic data structure that maintains a ((1-varepsilon ))-approximate maximum weight matching under (1) one-sided vertex deletions (with incident edges) and (2) one-sided vertex insertions (with incident edges sorted by weight) to the other side. The total time time used is (O(mvarepsilon ^{-1})), where m is the sum of the number of initially existing and inserted edges.
摘要 我们提出了一种使用乘法而非恒定权重更新的拍卖算法,以计算在具有 n 个顶点和 m 条边的双向图中的((1-varepsilon ))近似最大权重匹配(MWM)。-(O(mvarepsilon ^{-1}))的时间内计算出具有 n 个顶点和 m 条边的双向图中的近似最大权重匹配(MWM)。打败了 Duan 和 Pettie [JACM '14] 的最快近似算法的运行时间(O(mvarepsilon ^{-1}log varepsilon ^{-1})) 。我们的算法非常简单,而且可以扩展到给出一个动态数据结构来维护一个((1-varepsilon ))-近似最大权重匹配。-近似最大权重匹配的动态数据结构。(1)单边顶点删除(带入射边)和(2)单边顶点插入(带按权重排序的入射边)到另一方。所用总时间为 (O(mvarepsilon ^{-1}))其中,m 是最初存在的和插入的边的数量之和。
{"title":"Multiplicative auction algorithm for approximate maximum weight bipartite matching","authors":"","doi":"10.1007/s10107-024-02066-3","DOIUrl":"https://doi.org/10.1007/s10107-024-02066-3","url":null,"abstract":"<h3>Abstract</h3> <p>We present an <em>auction algorithm</em> using multiplicative instead of constant weight updates to compute a <span> <span>((1-varepsilon ))</span> </span>-approximate maximum weight matching (MWM) in a bipartite graph with <em>n</em> vertices and <em>m</em> edges in time <span> <span>(O(mvarepsilon ^{-1}))</span> </span>, beating the running time of the fastest known approximation algorithm of Duan and Pettie [JACM ’14] that runs in <span> <span>(O(mvarepsilon ^{-1}log varepsilon ^{-1}))</span> </span>. Our algorithm is very simple and it can be extended to give a dynamic data structure that maintains a <span> <span>((1-varepsilon ))</span> </span>-approximate maximum weight matching under (1) one-sided vertex deletions (with incident edges) and (2) one-sided vertex insertions (with incident edges sorted by weight) to the other side. The total time time used is <span> <span>(O(mvarepsilon ^{-1}))</span> </span>, where <em>m</em> is the sum of the number of initially existing and inserted edges.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"55 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140054991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-04DOI: 10.1007/s10107-024-02058-3
Eitan Levin, Joe Kileel, Nicolas Boumal
We develop new tools to study landscapes in nonconvex optimization. Given one optimization problem, we pair it with another by smoothly parametrizing the domain. This is either for practical purposes (e.g., to use smooth optimization algorithms with good guarantees) or for theoretical purposes (e.g., to reveal that the landscape satisfies a strict saddle property). In both cases, the central question is: how do the landscapes of the two problems relate? More precisely: how do desirable points such as local minima and critical points in one problem relate to those in the other problem? A key finding in this paper is that these relations are often determined by the parametrization itself, and are almost entirely independent of the cost function. Accordingly, we introduce a general framework to study parametrizations by their effect on landscapes. The framework enables us to obtain new guarantees for an array of problems, some of which were previously treated on a case-by-case basis in the literature. Applications include: optimizing low-rank matrices and tensors through factorizations; solving semidefinite programs via the Burer–Monteiro approach; training neural networks by optimizing their weights and biases; and quotienting out symmetries.
{"title":"The effect of smooth parametrizations on nonconvex optimization landscapes","authors":"Eitan Levin, Joe Kileel, Nicolas Boumal","doi":"10.1007/s10107-024-02058-3","DOIUrl":"https://doi.org/10.1007/s10107-024-02058-3","url":null,"abstract":"<p>We develop new tools to study landscapes in nonconvex optimization. Given one optimization problem, we pair it with another by smoothly parametrizing the domain. This is either for practical purposes (e.g., to use smooth optimization algorithms with good guarantees) or for theoretical purposes (e.g., to reveal that the landscape satisfies a strict saddle property). In both cases, the central question is: how do the landscapes of the two problems relate? More precisely: how do desirable points such as local minima and critical points in one problem relate to those in the other problem? A key finding in this paper is that these relations are often determined by the parametrization itself, and are almost entirely independent of the cost function. Accordingly, we introduce a general framework to study parametrizations by their effect on landscapes. The framework enables us to obtain new guarantees for an array of problems, some of which were previously treated on a case-by-case basis in the literature. Applications include: optimizing low-rank matrices and tensors through factorizations; solving semidefinite programs via the Burer–Monteiro approach; training neural networks by optimizing their weights and biases; and quotienting out symmetries.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"43 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140036042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-04DOI: 10.1007/s10107-024-02062-7
Pavel Dvurechensky, Mathias Staudigl
A key problem in mathematical imaging, signal processing and computational statistics is the minimization of non-convex objective functions that may be non-differentiable at the relative boundary of the feasible set. This paper proposes a new family of first- and second-order interior-point methods for non-convex optimization problems with linear and conic constraints, combining logarithmically homogeneous barriers with quadratic and cubic regularization respectively. Our approach is based on a potential-reduction mechanism and, under the Lipschitz continuity of the corresponding derivative with respect to the local barrier-induced norm, attains a suitably defined class of approximate first- or second-order KKT points with worst-case iteration complexity (O(varepsilon ^{-2})) (first-order) and (O(varepsilon ^{-3/2})) (second-order), respectively. Based on these findings, we develop new path-following schemes attaining the same complexity, modulo adjusting constants. These complexity bounds are known to be optimal in the unconstrained case, and our work shows that they are upper bounds in the case with complicated constraints as well. To the best of our knowledge, this work is the first which achieves these worst-case complexity bounds under such weak conditions for general conic constrained non-convex optimization problems.
{"title":"Hessian barrier algorithms for non-convex conic optimization","authors":"Pavel Dvurechensky, Mathias Staudigl","doi":"10.1007/s10107-024-02062-7","DOIUrl":"https://doi.org/10.1007/s10107-024-02062-7","url":null,"abstract":"<p>A key problem in mathematical imaging, signal processing and computational statistics is the minimization of non-convex objective functions that may be non-differentiable at the relative boundary of the feasible set. This paper proposes a new family of first- and second-order interior-point methods for non-convex optimization problems with linear and conic constraints, combining logarithmically homogeneous barriers with quadratic and cubic regularization respectively. Our approach is based on a potential-reduction mechanism and, under the Lipschitz continuity of the corresponding derivative with respect to the local barrier-induced norm, attains a suitably defined class of approximate first- or second-order KKT points with worst-case iteration complexity <span>(O(varepsilon ^{-2}))</span> (first-order) and <span>(O(varepsilon ^{-3/2}))</span> (second-order), respectively. Based on these findings, we develop new path-following schemes attaining the same complexity, modulo adjusting constants. These complexity bounds are known to be optimal in the unconstrained case, and our work shows that they are upper bounds in the case with complicated constraints as well. To the best of our knowledge, this work is the first which achieves these worst-case complexity bounds under such weak conditions for general conic constrained non-convex optimization problems.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"4 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140036039","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}