Pub Date : 2024-11-01Epub Date: 2023-02-22DOI: 10.1080/00324728.2023.2174266
Eric B Schneider, Sören Edvinsson, Kota Ogasawara
While there is strong evidence that maternal smallpox infection can cause foetal loss, it is not clear whether smallpox infections were a demographically important cause of stillbirths historically. In this paper, we use parish-level data from the Swedish Tabellverket data set for 1780-1839 to test the effect of smallpox on stillbirths quantitatively, analysing periods before and after the introduction of vaccination in 1802. We find that smallpox infection was not a major cause of stillbirths before 1820, because most women contracted smallpox as children and were therefore not susceptible during pregnancy. We do find a small, statistically significant effect of smallpox on stillbirths from 1820 to 1839, when waning immunity from vaccination put a greater share of pregnant women at risk of contracting smallpox. However, the reduced prevalence of smallpox in this period limited its impact on stillbirths. Thus, smallpox was not an important driver of historical stillbirth trends.
{"title":"Did smallpox cause stillbirths? Maternal smallpox infection, vaccination, and stillbirths in Sweden, 1780-1839.","authors":"Eric B Schneider, Sören Edvinsson, Kota Ogasawara","doi":"10.1080/00324728.2023.2174266","DOIUrl":"10.1080/00324728.2023.2174266","url":null,"abstract":"<p><p>While there is strong evidence that maternal smallpox infection can cause foetal loss, it is not clear whether smallpox infections were a demographically important cause of stillbirths historically. In this paper, we use parish-level data from the Swedish Tabellverket data set for 1780-1839 to test the effect of smallpox on stillbirths quantitatively, analysing periods before and after the introduction of vaccination in 1802. We find that smallpox infection was not a major cause of stillbirths before 1820, because most women contracted smallpox as children and were therefore not susceptible during pregnancy. We do find a small, statistically significant effect of smallpox on stillbirths from 1820 to 1839, when waning immunity from vaccination put a greater share of pregnant women at risk of contracting smallpox. However, the reduced prevalence of smallpox in this period limited its impact on stillbirths. Thus, smallpox was not an important driver of historical stillbirth trends.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"15 1","pages":"467-482"},"PeriodicalIF":1.5,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73348221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s10107-024-02136-6
Quentin Rebjock, Nicolas Boumal
Optimization algorithms can see their local convergence rates deteriorate when the Hessian at the optimum is singular. These singularities are inescapable when the optima are non-isolated. Yet, under the right circumstances, several algorithms preserve their favorable rates even when optima form a continuum (e.g., due to over-parameterization). This has been explained under various structural assumptions, including the Polyak–Łojasiewicz condition, Quadratic Growth and the Error Bound. We show that, for cost functions which are twice continuously differentiable ((textrm{C}^2)), those three (local) properties are equivalent. Moreover, we show they are equivalent to the Morse–Bott property, that is, local minima form differentiable submanifolds, and the Hessian of the cost function is positive definite along its normal directions. We leverage this insight to improve local convergence guarantees for safe-guarded Newton-type methods under any (hence all) of the above assumptions. First, for adaptive cubic regularization, we secure quadratic convergence even with inexact subproblem solvers. Second, for trust-region methods, we argue capture can fail with an exact subproblem solver, then proceed to show linear convergence with an inexact one (Cauchy steps).
{"title":"Fast convergence to non-isolated minima: four equivalent conditions for $${textrm{C}^{2}}$$ functions","authors":"Quentin Rebjock, Nicolas Boumal","doi":"10.1007/s10107-024-02136-6","DOIUrl":"https://doi.org/10.1007/s10107-024-02136-6","url":null,"abstract":"<p>Optimization algorithms can see their local convergence rates deteriorate when the Hessian at the optimum is singular. These singularities are inescapable when the optima are non-isolated. Yet, under the right circumstances, several algorithms preserve their favorable rates even when optima form a continuum (e.g., due to over-parameterization). This has been explained under various structural assumptions, including the Polyak–Łojasiewicz condition, Quadratic Growth and the Error Bound. We show that, for cost functions which are twice continuously differentiable (<span>(textrm{C}^2)</span>), those three (local) properties are equivalent. Moreover, we show they are equivalent to the Morse–Bott property, that is, local minima form differentiable submanifolds, and the Hessian of the cost function is positive definite along its normal directions. We leverage this insight to improve local convergence guarantees for safe-guarded Newton-type methods under any (hence all) of the above assumptions. First, for adaptive cubic regularization, we secure quadratic convergence even with inexact subproblem solvers. Second, for trust-region methods, we argue capture can fail with an exact subproblem solver, then proceed to show linear convergence with an inexact one (Cauchy steps).</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"29 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-17DOI: 10.1007/s10107-024-02137-5
Richard Y. Zhang
If a sparse semidefinite program (SDP), specified over (ntimes n) matrices and subject to m linear constraints, has an aggregate sparsity graph G with small treewidth, then chordal conversion will sometimes allow an interior-point method to solve the SDP in just (O(m+n)) time per-iteration, which is a significant speedup over the (varOmega (n^{3})) time per-iteration for a direct application of the interior-point method. Unfortunately, the speedup is not guaranteed by an O(1) treewidth in G that is independent of m and n, as a diagonal SDP would have treewidth zero but can still necessitate up to (varOmega (n^{3})) time per-iteration. Instead, we construct an extended aggregate sparsity graph (overline{G}supseteq G) by forcing each constraint matrix (A_{i}) to be its own clique in G. We prove that a small treewidth in (overline{G}) does indeed guarantee that chordal conversion will solve the SDP in (O(m+n)) time per-iteration, to (epsilon )-accuracy in at most (O(sqrt{m+n}log (1/epsilon ))) iterations. This sufficient condition covers many successful applications of chordal conversion, including the MAX-k-CUT relaxation, the Lovász theta problem, sensor network localization, polynomial optimization, and the AC optimal power flow relaxation, thus allowing theory to match practical experience.
如果一个在 (ntimes n) 矩阵上指定并受制于 m 个线性约束的稀疏半定式程序(SDP)有一个具有小树宽的总稀疏性图 G、那么和弦转换有时会让内点法在每次迭代时只需要 (O(m+n)) 时间就能求解 SDP,这比直接应用内点法每次迭代所需的(varOmega (n^{3})) 时间大大加快了速度。不幸的是,这种加速并不能通过 G 中与 m 和 n 无关的 O(1) 树状宽度来保证,因为对角 SDP 的树状宽度为零,但每次迭代仍然需要多达 (varOmega (n^{3})) 的时间。相反,我们通过强迫每个约束矩阵 (A_{i}) 成为 G 中自己的小块,来构建一个扩展的集合稀疏性图 (overline{G}supseteq G) 。我们证明,(overline{G})中的小树宽确实可以保证弦变换在每次迭代中以(O(m+n))时间求解 SDP,最多以(O(sqrt{m+n}log (1/epsilon ))) 次迭代达到(epsilon)精度。这个充分条件涵盖了和弦转换的许多成功应用,包括 MAX-k-CUT 松弛、Lovász theta 问题、传感器网络定位、多项式优化和交流最优功率流松弛,从而使理论与实践经验相匹配。
{"title":"Complexity of chordal conversion for sparse semidefinite programs with small treewidth","authors":"Richard Y. Zhang","doi":"10.1007/s10107-024-02137-5","DOIUrl":"https://doi.org/10.1007/s10107-024-02137-5","url":null,"abstract":"<p>If a sparse semidefinite program (SDP), specified over <span>(ntimes n)</span> matrices and subject to <i>m</i> linear constraints, has an aggregate sparsity graph <i>G</i> with small treewidth, then chordal conversion will sometimes allow an interior-point method to solve the SDP in just <span>(O(m+n))</span> time per-iteration, which is a significant speedup over the <span>(varOmega (n^{3}))</span> time per-iteration for a direct application of the interior-point method. Unfortunately, the speedup is not guaranteed by an <i>O</i>(1) treewidth in <i>G</i> that is independent of <i>m</i> and <i>n</i>, as a diagonal SDP would have treewidth zero but can still necessitate up to <span>(varOmega (n^{3}))</span> time per-iteration. Instead, we construct an extended aggregate sparsity graph <span>(overline{G}supseteq G)</span> by forcing each constraint matrix <span>(A_{i})</span> to be its own clique in <i>G</i>. We prove that a small treewidth in <span>(overline{G})</span> does indeed guarantee that chordal conversion will solve the SDP in <span>(O(m+n))</span> time per-iteration, to <span>(epsilon )</span>-accuracy in at most <span>(O(sqrt{m+n}log (1/epsilon )))</span> iterations. This sufficient condition covers many successful applications of chordal conversion, including the MAX-<i>k</i>-CUT relaxation, the Lovász theta problem, sensor network localization, polynomial optimization, and the AC optimal power flow relaxation, thus allowing theory to match practical experience.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"15 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-29DOI: 10.1007/s10107-024-02135-7
Christina Büsing, Timo Gersing, Arie M. C. A. Koster
Robust combinatorial optimization with budgeted uncertainty is one of the most popular approaches for integrating uncertainty into optimization problems. The existence of a compact reformulation for (mixed-integer) linear programs and positive complexity results give the impression that these problems are relatively easy to solve. However, the practical performance of the reformulation is quite poor when solving robust integer problems, in particular due to its weak linear relaxation. To overcome this issue, we propose procedures to derive new classes of valid inequalities for robust combinatorial optimization problems. For this, we recycle valid inequalities of the underlying deterministic problem such that the additional variables from the robust formulation are incorporated. The valid inequalities to be recycled may either be readily available model constraints or actual cutting planes, where we can benefit from decades of research on valid inequalities for classical optimization problems. We first demonstrate the strength of the inequalities theoretically, by proving that recycling yields a facet-defining inequality in many cases, even if the original valid inequality was not facet-defining. Afterwards, we show in an extensive computational study that using recycled inequalities can lead to a significant improvement of the computation time when solving robust optimization problems.
{"title":"Recycling valid inequalities for robust combinatorial optimization with budgeted uncertainty","authors":"Christina Büsing, Timo Gersing, Arie M. C. A. Koster","doi":"10.1007/s10107-024-02135-7","DOIUrl":"https://doi.org/10.1007/s10107-024-02135-7","url":null,"abstract":"<p>Robust combinatorial optimization with budgeted uncertainty is one of the most popular approaches for integrating uncertainty into optimization problems. The existence of a compact reformulation for (mixed-integer) linear programs and positive complexity results give the impression that these problems are relatively easy to solve. However, the practical performance of the reformulation is quite poor when solving robust integer problems, in particular due to its weak linear relaxation. To overcome this issue, we propose procedures to derive new classes of valid inequalities for robust combinatorial optimization problems. For this, we recycle valid inequalities of the underlying deterministic problem such that the additional variables from the robust formulation are incorporated. The valid inequalities to be recycled may either be readily available model constraints or actual cutting planes, where we can benefit from decades of research on valid inequalities for classical optimization problems. We first demonstrate the strength of the inequalities theoretically, by proving that recycling yields a facet-defining inequality in many cases, even if the original valid inequality was not facet-defining. Afterwards, we show in an extensive computational study that using recycled inequalities can lead to a significant improvement of the computation time when solving robust optimization problems.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"152 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-27DOI: 10.1007/s10107-024-02138-4
Sasila Ilandarideva, Anatoli Juditsky, Guanghui Lan, Tianjiao Li
We consider a class of stochastic smooth convex optimization problems under rather general assumptions on the noise in the stochastic gradient observation. As opposed to the classical problem setting in which the variance of noise is assumed to be uniformly bounded, herein we assume that the variance of stochastic gradients is related to the “sub-optimality” of the approximate solutions delivered by the algorithm. Such problems naturally arise in a variety of applications, in particular, in the well-known generalized linear regression problem in statistics. However, to the best of our knowledge, none of the existing stochastic approximation algorithms for solving this class of problems attain optimality in terms of the dependence on accuracy, problem parameters, and mini-batch size. We discuss two non-Euclidean accelerated stochastic approximation routines—stochastic accelerated gradient descent (SAGD) and stochastic gradient extrapolation (SGE)—which carry a particular duality relationship. We show that both SAGD and SGE, under appropriate conditions, achieve the optimal convergence rate, attaining the optimal iteration and sample complexities simultaneously. However, corresponding assumptions for the SGE algorithm are more general; they allow, for instance, for efficient application of the SGE to statistical estimation problems under heavy tail noises and discontinuous score functions. We also discuss the application of the SGE to problems satisfying quadratic growth conditions, and show how it can be used to recover sparse solutions. Finally, we report on some simulation experiments to illustrate numerical performance of our proposed algorithms in high-dimensional settings.
{"title":"Accelerated stochastic approximation with state-dependent noise","authors":"Sasila Ilandarideva, Anatoli Juditsky, Guanghui Lan, Tianjiao Li","doi":"10.1007/s10107-024-02138-4","DOIUrl":"https://doi.org/10.1007/s10107-024-02138-4","url":null,"abstract":"<p>We consider a class of stochastic smooth convex optimization problems under rather general assumptions on the noise in the stochastic gradient observation. As opposed to the classical problem setting in which the variance of noise is assumed to be uniformly bounded, herein we assume that the variance of stochastic gradients is related to the “sub-optimality” of the approximate solutions delivered by the algorithm. Such problems naturally arise in a variety of applications, in particular, in the well-known generalized linear regression problem in statistics. However, to the best of our knowledge, none of the existing stochastic approximation algorithms for solving this class of problems attain optimality in terms of the dependence on accuracy, problem parameters, and mini-batch size. We discuss two non-Euclidean accelerated stochastic approximation routines—stochastic accelerated gradient descent (SAGD) and stochastic gradient extrapolation (SGE)—which carry a particular duality relationship. We show that both SAGD and SGE, under appropriate conditions, achieve the optimal convergence rate, attaining the optimal iteration and sample complexities simultaneously. However, corresponding assumptions for the SGE algorithm are more general; they allow, for instance, for efficient application of the SGE to statistical estimation problems under heavy tail noises and discontinuous score functions. We also discuss the application of the SGE to problems satisfying quadratic growth conditions, and show how it can be used to recover sparse solutions. Finally, we report on some simulation experiments to illustrate numerical performance of our proposed algorithms in high-dimensional settings.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"62 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s10107-024-02133-9
Noah Weninger, Ricardo Fukasawa
We consider the bilevel knapsack problem with interdiction constraints, a fundamental bilevel integer programming problem which generalizes the 0–1 knapsack problem. In this problem, there are two knapsacks and n items. The objective is to select some items to pack into the first knapsack such that the maximum profit attainable from packing some of the remaining items into the second knapsack is minimized. We present a combinatorial branch-and-bound algorithm which outperforms the current state-of-the-art solution method in computational experiments for 99% of the instances reported in the literature. On many of the harder instances, our algorithm is orders of magnitude faster, which enabled it to solve 53 of the 72 previously unsolved instances. Our result relies fundamentally on a new dynamic programming algorithm which computes very strong lower bounds. This dynamic program solves a relaxation of the problem from bilevel to 2n-level where the items are processed in an online fashion. The relaxation is easier to solve but approximates the original problem surprisingly well in practice. We believe that this same technique may be useful for other interdiction problems.
{"title":"A fast combinatorial algorithm for the bilevel knapsack problem with interdiction constraints","authors":"Noah Weninger, Ricardo Fukasawa","doi":"10.1007/s10107-024-02133-9","DOIUrl":"https://doi.org/10.1007/s10107-024-02133-9","url":null,"abstract":"<p>We consider the bilevel knapsack problem with interdiction constraints, a fundamental bilevel integer programming problem which generalizes the 0–1 knapsack problem. In this problem, there are two knapsacks and <i>n</i> items. The objective is to select some items to pack into the first knapsack such that the maximum profit attainable from packing some of the remaining items into the second knapsack is minimized. We present a combinatorial branch-and-bound algorithm which outperforms the current state-of-the-art solution method in computational experiments for 99% of the instances reported in the literature. On many of the harder instances, our algorithm is orders of magnitude faster, which enabled it to solve 53 of the 72 previously unsolved instances. Our result relies fundamentally on a new dynamic programming algorithm which computes very strong lower bounds. This dynamic program solves a relaxation of the problem from bilevel to 2<i>n</i>-level where the items are processed in an online fashion. The relaxation is easier to solve but approximates the original problem surprisingly well in practice. We believe that this same technique may be useful for other interdiction problems.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"2 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s10107-024-02127-7
Shuvomoy Das Gupta, Robert M. Freund, Xu Andy Sun, Adrien Taylor
We propose a computer-assisted approach to the analysis of the worst-case convergence of nonlinear conjugate gradient methods (NCGMs). Those methods are known for their generally good empirical performances for large-scale optimization, while having relatively incomplete analyses. Using our computer-assisted approach, we establish novel complexity bounds for the Polak-Ribière-Polyak (PRP) and the Fletcher-Reeves (FR) NCGMs for smooth strongly convex minimization. In particular, we construct mathematical proofs that establish the first non-asymptotic convergence bound for FR (which is historically the first developed NCGM), and a much improved non-asymptotic convergence bound for PRP. Additionally, we provide simple adversarial examples on which these methods do not perform better than gradient descent with exact line search, leaving very little room for improvements on the same class of problems.
{"title":"Nonlinear conjugate gradient methods: worst-case convergence rates via computer-assisted analyses","authors":"Shuvomoy Das Gupta, Robert M. Freund, Xu Andy Sun, Adrien Taylor","doi":"10.1007/s10107-024-02127-7","DOIUrl":"https://doi.org/10.1007/s10107-024-02127-7","url":null,"abstract":"<p>We propose a computer-assisted approach to the analysis of the worst-case convergence of nonlinear conjugate gradient methods (NCGMs). Those methods are known for their generally good empirical performances for large-scale optimization, while having relatively incomplete analyses. Using our computer-assisted approach, we establish novel complexity bounds for the Polak-Ribière-Polyak (PRP) and the Fletcher-Reeves (FR) NCGMs for smooth strongly convex minimization. In particular, we construct mathematical proofs that establish the first non-asymptotic convergence bound for FR (which is historically the first developed NCGM), and a much improved non-asymptotic convergence bound for PRP. Additionally, we provide simple adversarial examples on which these methods do not perform better than gradient descent with exact line search, leaving very little room for improvements on the same class of problems.\u0000</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"32 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1007/s10107-024-02130-y
Lara Scavuzzo, Karen Aardal, Andrea Lodi, Neil Yorke-Smith
Mixed Integer Linear Programming (MILP) is a pillar of mathematical optimization that offers a powerful modeling language for a wide range of applications. The main engine for solving MILPs is the branch-and-bound algorithm. Adding to the enormous algorithmic progress in MILP solving of the past decades, in more recent years there has been an explosive development in the use of machine learning for enhancing all main tasks involved in the branch-and-bound algorithm. These include primal heuristics, branching, cutting planes, node selection and solver configuration decisions. This article presents a survey of such approaches, addressing the vision of integration of machine learning and mathematical optimization as complementary technologies, and how this integration can benefit MILP solving. In particular, we give detailed attention to machine learning algorithms that automatically optimize some metric of branch-and-bound efficiency. We also address appropriate MILP representations, benchmarks and software tools used in the context of applying learning algorithms.
{"title":"Machine learning augmented branch and bound for mixed integer linear programming","authors":"Lara Scavuzzo, Karen Aardal, Andrea Lodi, Neil Yorke-Smith","doi":"10.1007/s10107-024-02130-y","DOIUrl":"https://doi.org/10.1007/s10107-024-02130-y","url":null,"abstract":"<p>Mixed Integer Linear Programming (MILP) is a pillar of mathematical optimization that offers a powerful modeling language for a wide range of applications. The main engine for solving MILPs is the branch-and-bound algorithm. Adding to the enormous algorithmic progress in MILP solving of the past decades, in more recent years there has been an explosive development in the use of machine learning for enhancing all main tasks involved in the branch-and-bound algorithm. These include primal heuristics, branching, cutting planes, node selection and solver configuration decisions. This article presents a survey of such approaches, addressing the vision of integration of machine learning and mathematical optimization as complementary technologies, and how this integration can benefit MILP solving. In particular, we give detailed attention to machine learning algorithms that automatically optimize some metric of branch-and-bound efficiency. We also address appropriate MILP representations, benchmarks and software tools used in the context of applying learning algorithms.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"6 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-20DOI: 10.1007/s10107-024-02121-z
Matthew Brun, Tyler Perini, Saumya Sinha, Andrew J. Schaefer
This paper investigates the potential of Lagrangian relaxations to generate quality bounds on non-dominated images of multiobjective integer programs (MOIPs). Under some conditions on the relaxed constraints, we show that a set of Lagrangian relaxations can provide bounds that coincide with every bound generated by the convex hull relaxation. We also provide a guarantee of the relative quality of the Lagrangian bound at unsupported solutions. These results imply that, if the relaxed feasible region is bounded, some Lagrangian bounds will be strictly better than some convex hull bounds. We demonstrate that there exist Lagrangian multipliers which are sparse, satisfy a complementary slackness property, and generate tight relaxations at supported solutions. However, if all constraints are dualized, a relaxation can never be tight at an unsupported solution. These results characterize the strength of the Lagrangian dual at efficient solutions of an MOIP.
{"title":"On the strength of Lagrangian duality in multiobjective integer programming","authors":"Matthew Brun, Tyler Perini, Saumya Sinha, Andrew J. Schaefer","doi":"10.1007/s10107-024-02121-z","DOIUrl":"https://doi.org/10.1007/s10107-024-02121-z","url":null,"abstract":"<p>This paper investigates the potential of Lagrangian relaxations to generate quality bounds on non-dominated images of multiobjective integer programs (MOIPs). Under some conditions on the relaxed constraints, we show that a set of Lagrangian relaxations can provide bounds that coincide with every bound generated by the convex hull relaxation. We also provide a guarantee of the relative quality of the Lagrangian bound at unsupported solutions. These results imply that, if the relaxed feasible region is bounded, some Lagrangian bounds will be strictly better than some convex hull bounds. We demonstrate that there exist Lagrangian multipliers which are sparse, satisfy a complementary slackness property, and generate tight relaxations at supported solutions. However, if all constraints are dualized, a relaxation can never be tight at an unsupported solution. These results characterize the strength of the Lagrangian dual at efficient solutions of an MOIP.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"23 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-16DOI: 10.1007/s10107-024-02131-x
Taotao He, Siyue Liu, Mohit Tawarmalani
This paper develops a correspondence relating convex hulls of fractional functions with those of polynomial functions over the same domain. Using this result, we develop a number of new reformulations and relaxations for fractional programming problems. First, we relate (0mathord {-}1) problems involving a ratio of affine functions with the boolean quadric polytope, and use inequalities for the latter to develop tighter formulations for the former. Second, we derive a new formulation to optimize a ratio of quadratic functions over a polytope using copositive programming. Third, we show that univariate fractional functions can be convexified using moment hulls. Fourth, we develop a new hierarchy of relaxations that converges finitely to the simultaneous convex hull of a collection of ratios of affine functions of (0mathord {-}1) variables. Finally, we demonstrate theoretically and computationally that our techniques close a significant gap relative to state-of-the-art relaxations, require much less computational effort, and can solve larger problem instances.
{"title":"Convexification techniques for fractional programs","authors":"Taotao He, Siyue Liu, Mohit Tawarmalani","doi":"10.1007/s10107-024-02131-x","DOIUrl":"https://doi.org/10.1007/s10107-024-02131-x","url":null,"abstract":"<p>This paper develops a correspondence relating convex hulls of fractional functions with those of polynomial functions over the same domain. Using this result, we develop a number of new reformulations and relaxations for fractional programming problems. First, we relate <span>(0mathord {-}1)</span> problems involving a ratio of affine functions with the boolean quadric polytope, and use inequalities for the latter to develop tighter formulations for the former. Second, we derive a new formulation to optimize a ratio of quadratic functions over a polytope using copositive programming. Third, we show that univariate fractional functions can be convexified using moment hulls. Fourth, we develop a new hierarchy of relaxations that converges finitely to the simultaneous convex hull of a collection of ratios of affine functions of <span>(0mathord {-}1)</span> variables. Finally, we demonstrate theoretically and computationally that our techniques close a significant gap relative to state-of-the-art relaxations, require much less computational effort, and can solve larger problem instances.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"33 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142183167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}