Pub Date : 2024-07-22DOI: 10.1007/s10107-024-02110-2
Wenqing Ouyang, Andre Milzarek
We propose a novel trust region method for solving a class of nonsmooth, nonconvex composite-type optimization problems. The approach embeds inexact semismooth Newton steps for finding zeros of a normal map-based stationarity measure for the problem in a trust region framework. Based on a new merit function and acceptance mechanism, global convergence and transition to fast local q-superlinear convergence are established under standard conditions. In addition, we verify that the proposed trust region globalization is compatible with the Kurdyka–Łojasiewicz inequality yielding finer convergence results. Experiments on sparse logistic regression, image compression, and a constrained log-determinant problem illustrate the efficiency of the proposed algorithm.
{"title":"A trust region-type normal map-based semismooth Newton method for nonsmooth nonconvex composite optimization","authors":"Wenqing Ouyang, Andre Milzarek","doi":"10.1007/s10107-024-02110-2","DOIUrl":"https://doi.org/10.1007/s10107-024-02110-2","url":null,"abstract":"<p>We propose a novel trust region method for solving a class of nonsmooth, nonconvex composite-type optimization problems. The approach embeds inexact semismooth Newton steps for finding zeros of a normal map-based stationarity measure for the problem in a trust region framework. Based on a new merit function and acceptance mechanism, global convergence and transition to fast local q-superlinear convergence are established under standard conditions. In addition, we verify that the proposed trust region globalization is compatible with the Kurdyka–Łojasiewicz inequality yielding finer convergence results. Experiments on sparse logistic regression, image compression, and a constrained log-determinant problem illustrate the efficiency of the proposed algorithm.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"4 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141740860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-22DOI: 10.1007/s10107-024-02118-8
Sven Jäger, Guillaume Sagnol, Daniel Schmidt genannt Waldschmidt, Philipp Warode
We study kill-and-restart and preemptive strategies for the fundamental scheduling problem of minimizing the sum of weighted completion times on a single machine in the non-clairvoyant setting. First, we show a lower bound of 3 for any deterministic non-clairvoyant kill-and-restart strategy. Then, we give for any (b > 1) a tight analysis for the natural b-scaling kill-and-restart strategy as well as for a randomized variant of it. In particular, we show a competitive ratio of ((1+3sqrt{3})approx 6.197) for the deterministic and of (approx 3.032) for the randomized strategy, by making use of the largest eigenvalue of a Toeplitz matrix. In addition, we show that the preemptive Weighted Shortest Elapsed Time First (WSETF) rule is 2-competitive when jobs are released online, matching the lower bound for the unit weight case with trivial release dates for any non-clairvoyant algorithm. Using this result as well as the competitiveness of round-robin for multiple machines, we prove performance guarantees smaller than 10 for adaptions of the b-scaling strategy to online release dates and unweighted jobs on identical parallel machines.
{"title":"Competitive kill-and-restart and preemptive strategies for non-clairvoyant scheduling","authors":"Sven Jäger, Guillaume Sagnol, Daniel Schmidt genannt Waldschmidt, Philipp Warode","doi":"10.1007/s10107-024-02118-8","DOIUrl":"https://doi.org/10.1007/s10107-024-02118-8","url":null,"abstract":"<p>We study kill-and-restart and preemptive strategies for the fundamental scheduling problem of minimizing the sum of weighted completion times on a single machine in the non-clairvoyant setting. First, we show a lower bound of 3 for any deterministic non-clairvoyant kill-and-restart strategy. Then, we give for any <span>(b > 1)</span> a tight analysis for the natural <i>b</i>-scaling kill-and-restart strategy as well as for a randomized variant of it. In particular, we show a competitive ratio of <span>((1+3sqrt{3})approx 6.197)</span> for the deterministic and of <span>(approx 3.032)</span> for the randomized strategy, by making use of the largest eigenvalue of a Toeplitz matrix. In addition, we show that the preemptive Weighted Shortest Elapsed Time First (WSETF) rule is 2-competitive when jobs are released online, matching the lower bound for the unit weight case with trivial release dates for any non-clairvoyant algorithm. Using this result as well as the competitiveness of round-robin for multiple machines, we prove performance guarantees smaller than 10 for adaptions of the <i>b</i>-scaling strategy to online release dates and unweighted jobs on identical parallel machines.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"26 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141746019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-17DOI: 10.1007/s10107-024-02109-9
Haihao Lu, Jinwen Yang
We study the convergence behaviors of primal–dual hybrid gradient (PDHG) for solving linear programming (LP). PDHG is the base algorithm of a new general-purpose first-order method LP solver, PDLP, which aims to scale up LP by taking advantage of modern computing architectures. Despite its numerical success, the theoretical understanding of PDHG for LP is still very limited; the previous complexity result relies on the global Hoffman constant of the KKT system, which is known to be very loose and uninformative. In this work, we aim to develop a fundamental understanding of the convergence behaviors of PDHG for LP and to develop a refined complexity rate that does not rely on the global Hoffman constant. We show that there are two major stages of PDHG for LP: in Stage I, PDHG identifies active variables and the length of the first stage is driven by a certain quantity which measures how close the non-degeneracy part of the LP instance is to degeneracy; in Stage II, PDHG effectively solves a homogeneous linear inequality system, and the complexity of the second stage is driven by a well-behaved local sharpness constant of the system. This finding is closely related to the concept of partial smoothness in non-smooth optimization, and it is the first complexity result of finite time identification without the non-degeneracy assumption. An interesting implication of our results is that degeneracy itself does not slow down the convergence of PDHG for LP, but near-degeneracy does.
{"title":"On the geometry and refined rate of primal–dual hybrid gradient for linear programming","authors":"Haihao Lu, Jinwen Yang","doi":"10.1007/s10107-024-02109-9","DOIUrl":"https://doi.org/10.1007/s10107-024-02109-9","url":null,"abstract":"<p>We study the convergence behaviors of primal–dual hybrid gradient (PDHG) for solving linear programming (LP). PDHG is the base algorithm of a new general-purpose first-order method LP solver, PDLP, which aims to scale up LP by taking advantage of modern computing architectures. Despite its numerical success, the theoretical understanding of PDHG for LP is still very limited; the previous complexity result relies on the global Hoffman constant of the KKT system, which is known to be very loose and uninformative. In this work, we aim to develop a fundamental understanding of the convergence behaviors of PDHG for LP and to develop a refined complexity rate that does not rely on the global Hoffman constant. We show that there are two major stages of PDHG for LP: in Stage I, PDHG identifies active variables and the length of the first stage is driven by a certain quantity which measures how close the non-degeneracy part of the LP instance is to degeneracy; in Stage II, PDHG effectively solves a homogeneous linear inequality system, and the complexity of the second stage is driven by a well-behaved local sharpness constant of the system. This finding is closely related to the concept of partial smoothness in non-smooth optimization, and it is the first complexity result of finite time identification without the non-degeneracy assumption. An interesting implication of our results is that degeneracy itself does not slow down the convergence of PDHG for LP, but near-degeneracy does.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"30 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141717803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-17DOI: 10.1007/s10107-024-02114-y
Jian Hu, Dali Zhang, Huifu Xu, Sainan Zhang
Utility preference robust optimization (PRO) has recently been proposed to deal with optimal decision-making problems where the decision maker’s (DM’s) preference over gains and losses is ambiguous. In this paper, we take a step further to investigate the case that the DM’s preference is random. We propose to use a random utility function to describe the DM’s preference and develop distributional utility preference robust optimization (DUPRO) models when the distribution of the random utility function is ambiguous. We concentrate on data-driven problems where samples of the random parameters are obtainable but the sample size may be relatively small. In the case when the random utility functions are of piecewise linear structure, we propose a bootstrap method to construct the ambiguity set and demonstrate how the resulting DUPRO can be solved by a mixed-integer linear program. The piecewise linear structure is versatile in its ability to incorporate classical non-parametric utility assessment methods into the sample generation of a random utility function. Next, we expand the proposed DUPRO models and computational schemes to address general cases where the random utility functions are not necessarily piecewise linear. We show how the DUPRO models with piecewise linear random utility functions can serve as approximations for the DUPRO models with general random utility functions and allow us to quantify the approximation errors. Finally, we carry out some performance studies of the proposed bootstrap-based DUPRO model and report the preliminary numerical test results. This paper is the first attempt to use distributionally robust optimization methods for PRO problems.
{"title":"Distributional utility preference robust optimization models in multi-attribute decision making","authors":"Jian Hu, Dali Zhang, Huifu Xu, Sainan Zhang","doi":"10.1007/s10107-024-02114-y","DOIUrl":"https://doi.org/10.1007/s10107-024-02114-y","url":null,"abstract":"<p>Utility preference robust optimization (PRO) has recently been proposed to deal with optimal decision-making problems where the decision maker’s (DM’s) preference over gains and losses is ambiguous. In this paper, we take a step further to investigate the case that the DM’s preference is random. We propose to use a random utility function to describe the DM’s preference and develop distributional utility preference robust optimization (DUPRO) models when the distribution of the random utility function is ambiguous. We concentrate on data-driven problems where samples of the random parameters are obtainable but the sample size may be relatively small. In the case when the random utility functions are of piecewise linear structure, we propose a bootstrap method to construct the ambiguity set and demonstrate how the resulting DUPRO can be solved by a mixed-integer linear program. The piecewise linear structure is versatile in its ability to incorporate classical non-parametric utility assessment methods into the sample generation of a random utility function. Next, we expand the proposed DUPRO models and computational schemes to address general cases where the random utility functions are not necessarily piecewise linear. We show how the DUPRO models with piecewise linear random utility functions can serve as approximations for the DUPRO models with general random utility functions and allow us to quantify the approximation errors. Finally, we carry out some performance studies of the proposed bootstrap-based DUPRO model and report the preliminary numerical test results. This paper is the first attempt to use distributionally robust optimization methods for PRO problems.\u0000</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"31 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141717805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-17DOI: 10.1007/s10107-024-02120-0
Alex L. Wang, Rujun Jiang
A set of quadratic forms is simultaneously diagonalizable via congruence (SDC) if there exists a basis under which each of the quadratic forms is diagonal. This property appears naturally when analyzing quadratically constrained quadratic programs (QCQPs) and has important implications in globally solving such problems using branch-and-bound methods. This paper extends the reach of the SDC property by studying two new weaker notions of simultaneous diagonalizability. Specifically, we say that a set of quadratic forms is almost SDC (ASDC) if it is the limit of SDC sets and d-restricted SDC (d-RSDC) if it is the restriction of an SDC set in up to d-many additional dimensions. In the context of QCQPs, these properties correspond to problems that may be diagonalized after arbitrarily small perturbations or after the introduction of d additional variables. Our main contributions are complete characterizations of the ASDC pairs and nonsingular triples of symmetric matrices, as well as a sufficient condition for the 1-RSDC property for pairs of symmetric matrices. Surprisingly, we show that every singular symmetric pair is ASDC and that almost every symmetric pair is 1-RSDC. We accompany our theoretical results with preliminary numerical experiments applying these constructions to solve QCQPs within branch-and-bound schemes.
如果存在这样一个基础,即每个二次型都是对角的,那么一组二次型就同时可通过全等对角(SDC)。这一性质在分析二次受限二次方程程序(QCQPs)时自然出现,并对使用分支约束法全局求解此类问题具有重要意义。本文通过研究两个新的较弱的同时对角化概念,扩展了 SDC 特性的范围。具体来说,如果一个二次型集合是 SDC 集合的极限,我们就说它几乎是 SDC (ASDC);如果它是 SDC 集合在多达 d 个额外维度上的限制,我们就说它是 d 限制 SDC (d-RSDC)。在 QCQPs 的背景下,这些性质对应于经过任意小的扰动或引入 d 个额外变量后可以对角化的问题。我们的主要贡献是完整描述了对称矩阵的 ASDC 对和非奇异三元组,以及对称矩阵对的 1-RSDC 属性的充分条件。令人惊讶的是,我们证明了每个奇异对称对都是 ASDC,而且几乎每个对称对都是 1-RSDC。在得出理论结果的同时,我们还进行了初步的数值实验,将这些构造应用于在分支与边界方案中求解 QCQP。
{"title":"New notions of simultaneous diagonalizability of quadratic forms with applications to QCQPs","authors":"Alex L. Wang, Rujun Jiang","doi":"10.1007/s10107-024-02120-0","DOIUrl":"https://doi.org/10.1007/s10107-024-02120-0","url":null,"abstract":"<p>A set of quadratic forms is simultaneously diagonalizable via congruence (SDC) if there exists a basis under which each of the quadratic forms is diagonal. This property appears naturally when analyzing quadratically constrained quadratic programs (QCQPs) and has important implications in globally solving such problems using branch-and-bound methods. This paper extends the reach of the SDC property by studying two new weaker notions of simultaneous diagonalizability. Specifically, we say that a set of quadratic forms is almost SDC (ASDC) if it is the limit of SDC sets and <i>d</i>-restricted SDC (<i>d</i>-RSDC) if it is the restriction of an SDC set in up to <i>d</i>-many additional dimensions. In the context of QCQPs, these properties correspond to problems that may be diagonalized after arbitrarily small perturbations or after the introduction of <i>d</i> additional variables. Our main contributions are complete characterizations of the ASDC pairs and nonsingular triples of symmetric matrices, as well as a sufficient condition for the 1-RSDC property for pairs of symmetric matrices. Surprisingly, we show that <i>every</i> singular symmetric pair is ASDC and that <i>almost every</i> symmetric pair is 1-RSDC. We accompany our theoretical results with preliminary numerical experiments applying these constructions to solve QCQPs within branch-and-bound schemes.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141717804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The reformulation–linearization technique (RLT) is a prominent approach to constructing tight linear relaxations of non-convex continuous and mixed-integer optimization problems. The goal of this paper is to extend the applicability and improve the performance of RLT for bilinear product relations. First, we present a method for detecting bilinear product relations implicitly contained in mixed-integer linear programs, which is based on analyzing linear constraints with binary variables, thus enabling the application of bilinear RLT to a new class of problems. Strategies for filtering product relations are discussed and tested. Our second contribution addresses the high computational cost of RLT cut separation, which presents one of the major difficulties in applying RLT efficiently in practice. We propose a new RLT cutting plane separation algorithm which identifies combinations of linear constraints and bound factors that are expected to yield an inequality that is violated by the current relaxation solution. This algorithm is applicable to RLT cuts generated for all types of bilinear terms, including but not limited to the detected implicit products. A detailed computational study based on independent implementations in two solvers evaluates the performance impact of the proposed methods.
{"title":"Efficient separation of RLT cuts for implicit and explicit bilinear terms","authors":"Ksenia Bestuzheva, Ambros Gleixner, Tobias Achterberg","doi":"10.1007/s10107-024-02104-0","DOIUrl":"https://doi.org/10.1007/s10107-024-02104-0","url":null,"abstract":"<p>The reformulation–linearization technique (RLT) is a prominent approach to constructing tight linear relaxations of non-convex continuous and mixed-integer optimization problems. The goal of this paper is to extend the applicability and improve the performance of RLT for bilinear product relations. First, we present a method for detecting bilinear product relations implicitly contained in mixed-integer linear programs, which is based on analyzing linear constraints with binary variables, thus enabling the application of bilinear RLT to a new class of problems. Strategies for filtering product relations are discussed and tested. Our second contribution addresses the high computational cost of RLT cut separation, which presents one of the major difficulties in applying RLT efficiently in practice. We propose a new RLT cutting plane separation algorithm which identifies combinations of linear constraints and bound factors that are expected to yield an inequality that is violated by the current relaxation solution. This algorithm is applicable to RLT cuts generated for all types of bilinear terms, including but not limited to the detected implicit products. A detailed computational study based on independent implementations in two solvers evaluates the performance impact of the proposed methods.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"26 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141722007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-10DOI: 10.1007/s10107-024-02119-7
Anupam Gupta, Benjamin Moseley, Rudy Zhou
This paper considers approximation algorithms for generalized k-median problems. These problems can be informally described as k-median with a constant number of extra constraints, and includes k-median with outliers, and knapsack median. Our first contribution is a pseudo-approximation algorithm for generalized k-median that outputs a 6.387-approximate solution, with a constant number of fractional variables. The algorithm builds on the iterative rounding framework introduced by Krishnaswamy, Li, and Sandeep for k-median with outliers as reported (Krishnaswamy et al. in: Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, 2018). The main technical innovation is allowing richer constraint sets in the iterative rounding and using the structure of the resulting extreme points. Using our pseudo-approximation algorithm, we give improved approximation algorithms for k-median with outliers and knapsack median. This involves combining our pseudo-approximation with pre- and post-processing steps to round a constant number of fractional variables at a small increase in cost. Our algorithms achieve approximation ratios (6.994 + epsilon ) and (6.387 + epsilon ) for k-median with outliers and knapsack median, respectively. These improve on the best-known approximation ratio (7.081 + epsilon ) for both problems as reported (Krishnaswamy et al. in: Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, 2018).
本文探讨了广义 k 中值问题的近似算法。这些问题可以被非正式地描述为带有恒定数量额外约束的 k-中值问题,包括带有异常值的 k-中值问题和 knapsack 中值问题。我们的第一个贡献是针对广义 k 中值问题提出了一种伪近似算法,它能在分数变量数量不变的情况下输出 6.387 近似解。该算法建立在 Krishnaswamy、Li 和 Sandeep 针对有离群值的 k-median 引入的迭代舍入框架基础上(Krishnaswamy et al:第 50 届 ACM SIGACT 计算理论年度研讨会论文集,2018 年)。主要的技术创新在于允许在迭代舍入中使用更丰富的约束集,并使用由此产生的极值点结构。利用我们的伪逼近算法,我们给出了带离群值的 k-median 和 knapsack median 的改进逼近算法。这包括将我们的伪逼近算法与前处理和后处理步骤相结合,以较小的成本增加对一定数量的小数变量进行舍入。我们的算法分别实现了 k-median with outliers 和 knapsack median 的近似率(6.994 + epsilon )和(6.387 + epsilon )。对于这两个问题,这些近似比(7.081 + epsilon )都有所提高(Krishnaswamy et al:第 50 届 ACM SIGACT 计算理论年度研讨会论文集,2018 年)。
{"title":"Structural iterative rounding for generalized k-median problems","authors":"Anupam Gupta, Benjamin Moseley, Rudy Zhou","doi":"10.1007/s10107-024-02119-7","DOIUrl":"https://doi.org/10.1007/s10107-024-02119-7","url":null,"abstract":"<p>This paper considers approximation algorithms for generalized <i>k</i>-median problems. These problems can be informally described as <i>k</i>-median with a constant number of extra constraints, and includes <i>k</i>-median with outliers, and knapsack median. Our first contribution is a pseudo-approximation algorithm for generalized <i>k</i>-median that outputs a 6.387-approximate solution, with a constant number of fractional variables. The algorithm builds on the iterative rounding framework introduced by Krishnaswamy, Li, and Sandeep for <i>k</i>-median with outliers as reported (Krishnaswamy et al. in: Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, 2018). The main technical innovation is allowing richer constraint sets in the iterative rounding and using the structure of the resulting extreme points. Using our pseudo-approximation algorithm, we give improved approximation algorithms for <i>k</i>-median with outliers and knapsack median. This involves combining our pseudo-approximation with pre- and post-processing steps to round a constant number of fractional variables at a small increase in cost. Our algorithms achieve approximation ratios <span>(6.994 + epsilon )</span> and <span>(6.387 + epsilon )</span> for <i>k</i>-median with outliers and knapsack median, respectively. These improve on the best-known approximation ratio <span>(7.081 + epsilon )</span> for both problems as reported (Krishnaswamy et al. in: Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, 2018).</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"37 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141567652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-08DOI: 10.1007/s10107-024-02107-x
Daniel Dadush, Zhuan Khye Koh, Bento Natura, László A. Végh
We study the circuit diameter of polyhedra, introduced by Borgwardt, Finhold, and Hemmecke (SIAM J. Discrete Math. 29(1), 113–121 (2015)) as a relaxation of the combinatorial diameter. We show that the circuit diameter of a system ({xin mathbb {R}^n:, Ax=b,, mathbb {0}le xle u}) for (Ain mathbb {R}^{mtimes n}) is bounded by (O(mmin {m, n - m}log (m+kappa _A)+nlog n)), where (kappa _A) is the circuit imbalance measure of the constraint matrix. This yields a strongly polynomial circuit diameter bound if e.g., all entries of A have polynomially bounded encoding length in n. Further, we present circuit augmentation algorithms for LPs using the minimum-ratio circuit cancelling rule. Even though the standard minimum-ratio circuit cancelling algorithm is not finite in general, our variant can solve an LP in (O(mn^2log (n+kappa _A))) augmentation steps.
我们研究由 Borgwardt、Finhold 和 Hemmecke(SIAM J. Discrete Math.29(1), 113-121 (2015))作为组合直径的放松而引入的。我们证明了一个系统的电路直径({xin mathbb {R}^n:Ax=b,, mathbb {0}le xle u}) for (Ain mathbb {R}^{mtimes n}) is bounded by (O(mmin {m, n - m}log (m+kappa _A)+nlog n)), where (kappa _A) is the circuit imbalance measure of the constraint matrix.如果 A 的所有条目在 n 中的编码长度都是多项式约束的,那么这就产生了一个强多项式电路直径约束。尽管标准的最小比值电路消除算法在一般情况下不是有限的,但我们的变体可以在(O(mn^2log (n+kappa _A))增强步骤内解决 LP。
{"title":"On circuit diameter bounds via circuit imbalances","authors":"Daniel Dadush, Zhuan Khye Koh, Bento Natura, László A. Végh","doi":"10.1007/s10107-024-02107-x","DOIUrl":"https://doi.org/10.1007/s10107-024-02107-x","url":null,"abstract":"<p>We study the circuit diameter of polyhedra, introduced by Borgwardt, Finhold, and Hemmecke (SIAM J. Discrete Math. <b>29</b>(1), 113–121 (2015)) as a relaxation of the combinatorial diameter. We show that the circuit diameter of a system <span>({xin mathbb {R}^n:, Ax=b,, mathbb {0}le xle u})</span> for <span>(Ain mathbb {R}^{mtimes n})</span> is bounded by <span>(O(mmin {m, n - m}log (m+kappa _A)+nlog n))</span>, where <span>(kappa _A)</span> is the circuit imbalance measure of the constraint matrix. This yields a strongly polynomial circuit diameter bound if e.g., all entries of <i>A</i> have polynomially bounded encoding length in <i>n</i>. Further, we present circuit augmentation algorithms for LPs using the minimum-ratio circuit cancelling rule. Even though the standard minimum-ratio circuit cancelling algorithm is not finite in general, our variant can solve an LP in <span>(O(mn^2log (n+kappa _A)))</span> augmentation steps.\u0000</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"25 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141567653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-08DOI: 10.1007/s10107-024-02112-0
Antonia Chmiela, Gonzalo Muñoz, Felipe Serrano
Using the recently proposed maximal quadratic-free sets and the well-known monoidal strengthening procedure, we show how to improve intersection cuts for quadratically-constrained optimization problems by exploiting integrality requirements. We provide an explicit construction that allows an efficient implementation of the strengthened cuts along with computational results showing their improvements over the standard intersection cuts. We also show that, in our setting, there is unique lifting which implies that our strengthening procedure is generating the best possible cut coefficients for the integer variables.
{"title":"Monoidal strengthening and unique lifting in MIQCPs","authors":"Antonia Chmiela, Gonzalo Muñoz, Felipe Serrano","doi":"10.1007/s10107-024-02112-0","DOIUrl":"https://doi.org/10.1007/s10107-024-02112-0","url":null,"abstract":"<p>Using the recently proposed maximal quadratic-free sets and the well-known monoidal strengthening procedure, we show how to improve intersection cuts for quadratically-constrained optimization problems by exploiting integrality requirements. We provide an explicit construction that allows an efficient implementation of the strengthened cuts along with computational results showing their improvements over the standard intersection cuts. We also show that, in our setting, there is <i>unique lifting</i> which implies that our strengthening procedure is generating the best possible cut coefficients for the integer variables.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"25 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141567656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-05DOI: 10.1007/s10107-024-02090-3
Zhe Zhang, Guanghui Lan
Recently, convex nested stochastic composite optimization (NSCO) has received considerable interest for its applications in reinforcement learning and risk-averse optimization. However, existing NSCO algorithms have worse stochastic oracle complexities, by orders of magnitude, than those for simpler stochastic optimization problems without nested structures. Additionally, these algorithms require all outer-layer functions to be smooth, a condition violated by some important applications. This raises a question regarding whether the nested composition make stochastic optimization more difficult in terms of oracle complexity. In this paper, we answer the question by developing order-optimal algorithms for convex NSCO problems constructed from an arbitrary composition of smooth, structured non-smooth, and general non-smooth layer functions. When all outer-layer functions are smooth, we propose a stochastic sequential dual (SSD) method to achieve an oracle complexity of (mathcal {O}(1/epsilon ^2)) (resp., (mathcal {O}(1/epsilon ))) when the problem is convex (resp., strongly convex). If any outer-layer function is non-smooth, we propose a non-smooth stochastic sequential dual (nSSD) method to achieve an (mathcal {O}(1/epsilon ^2)) oracle complexity. We provide a lower complexity bound to show the latter (mathcal {O}(1/epsilon ^2)) complexity to be unimprovable, even under a strongly convex setting. All these complexity results seem to be new in the literature, and they indicate that convex NSCO problems have the same order of oracle complexity as problems without nested composition, except in the strongly convex and outer non-smooth cases.
{"title":"Optimal methods for convex nested stochastic composite optimization","authors":"Zhe Zhang, Guanghui Lan","doi":"10.1007/s10107-024-02090-3","DOIUrl":"https://doi.org/10.1007/s10107-024-02090-3","url":null,"abstract":"<p>Recently, convex nested stochastic composite optimization (NSCO) has received considerable interest for its applications in reinforcement learning and risk-averse optimization. However, existing NSCO algorithms have worse stochastic oracle complexities, by orders of magnitude, than those for simpler stochastic optimization problems without nested structures. Additionally, these algorithms require all outer-layer functions to be smooth, a condition violated by some important applications. This raises a question regarding whether the nested composition make stochastic optimization more difficult in terms of oracle complexity. In this paper, we answer the question by developing order-optimal algorithms for convex NSCO problems constructed from an arbitrary composition of smooth, structured non-smooth, and general non-smooth layer functions. When all outer-layer functions are smooth, we propose a stochastic sequential dual (SSD) method to achieve an oracle complexity of <span>(mathcal {O}(1/epsilon ^2))</span> (resp., <span>(mathcal {O}(1/epsilon ))</span>) when the problem is convex (resp., strongly convex). If any outer-layer function is non-smooth, we propose a non-smooth stochastic sequential dual (nSSD) method to achieve an <span>(mathcal {O}(1/epsilon ^2))</span> oracle complexity. We provide a lower complexity bound to show the latter <span>(mathcal {O}(1/epsilon ^2))</span> complexity to be unimprovable, even under a strongly convex setting. All these complexity results seem to be new in the literature, and they indicate that convex NSCO problems have the same order of oracle complexity as problems without nested composition, except in the strongly convex and outer non-smooth cases.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"59 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141547945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}