Pub Date : 2024-01-04DOI: 10.1007/s10107-023-02044-1
Vincent Cohen-Addad, Tobias Mömke, Victor Verdugo
In the non-uniform sparsest cut problem, we are given a supply graph G and a demand graph D, both with the same set of nodes V. The goal is to find a cut of V that minimizes the ratio of the total capacity on the edges of G crossing the cut over the total demand of the crossing edges of D. In this work, we study the non-uniform sparsest cut problem for supply graphs with bounded treewidth k. For this case, Gupta et al. (ACM STOC, 2013) obtained a 2-approximation with polynomial running time for fixed k, and it remained open the question of whether there exists a c-approximation algorithm for a constant c independent of k, that runs in (textsf{FPT}) time. We answer this question in the affirmative. We design a 2-approximation algorithm for the non-uniform sparsest cut with bounded treewidth supply graphs that runs in (textsf{FPT}) time, when parameterized by the treewidth. Our algorithm is based on rounding the optimal solution of a linear programming relaxation inspired by the Sherali-Adams hierarchy. In contrast to the classic Sherali-Adams approach, we construct a relaxation driven by a tree decomposition of the supply graph by including a carefully chosen set of lifting variables and constraints to encode information of subsets of nodes with super-constant size, and at the same time we have a sufficiently small linear program that can be solved in (textsf{FPT}) time.
在非均匀最疏剪切问题中,我们给定了一个供应图 G 和一个需求图 D,两者都有相同的节点集 V。我们的目标是找到 V 的一个剪切点,该剪切点能使 G 的交叉边上的总容量与 D 的交叉边上的总需求之比最小化。在这项工作中,我们将研究具有有界树宽 k 的供应图的非均匀最疏剪切问题。对于这种情况,Gupta 等人(ACM STOC,2013 年)在固定 k 的情况下获得了运行时间为多项式的 2-approximation 算法,而对于与 k 无关的常数 c,是否存在一种运行时间为 (textsf{FPT}) 的 c-approximation 算法,这个问题仍然悬而未决。我们的回答是肯定的。我们为具有有界树宽的非均匀最疏剪切供应图设计了一种 2-approximation 算法,当以树宽为参数时,该算法能在(textsf{FPT}) 时间内运行。我们的算法基于对受 Sherali-Adams 层次结构启发的线性规划松弛的最优解进行舍入。与经典的 Sherali-Adams 方法不同的是,我们构建了一种由供应图的树形分解驱动的松弛,包括精心选择的一组提升变量和约束条件,以编码具有超常大小的节点子集的信息,同时我们有一个足够小的线性规划,可以在 (textsf{FPT})时间内求解。
{"title":"A 2-approximation for the bounded treewidth sparsest cut problem in $$textsf{FPT}$$ Time","authors":"Vincent Cohen-Addad, Tobias Mömke, Victor Verdugo","doi":"10.1007/s10107-023-02044-1","DOIUrl":"https://doi.org/10.1007/s10107-023-02044-1","url":null,"abstract":"<p>In the non-uniform sparsest cut problem, we are given a supply graph <i>G</i> and a demand graph <i>D</i>, both with the same set of nodes <i>V</i>. The goal is to find a cut of <i>V</i> that minimizes the ratio of the total capacity on the edges of <i>G</i> crossing the cut over the total demand of the crossing edges of <i>D</i>. In this work, we study the non-uniform sparsest cut problem for supply graphs with bounded treewidth <i>k</i>. For this case, Gupta et al. (ACM STOC, 2013) obtained a 2-approximation with polynomial running time for fixed <i>k</i>, and it remained open the question of whether there exists a <i>c</i>-approximation algorithm for a constant <i>c</i> independent of <i>k</i>, that runs in <span>(textsf{FPT})</span> time. We answer this question in the affirmative. We design a 2-approximation algorithm for the non-uniform sparsest cut with bounded treewidth supply graphs that runs in <span>(textsf{FPT})</span> time, when parameterized by the treewidth. Our algorithm is based on rounding the optimal solution of a linear programming relaxation inspired by the Sherali-Adams hierarchy. In contrast to the classic Sherali-Adams approach, we construct a relaxation driven by a tree decomposition of the supply graph by including a carefully chosen set of lifting variables and constraints to encode information of subsets of nodes with super-constant size, and at the same time we have a sufficiently small linear program that can be solved in <span>(textsf{FPT})</span> time.\u0000</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"21 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139105285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-04DOI: 10.1007/s10107-023-02046-z
Abstract
In this paper, a special case of the generalized 4-block n-fold IPs is investigated, where (B_i=B) and B has a rank at most 1. Such IPs, called almost combinatorial 4-block n-fold IPs, include the generalized n-fold IPs as a subcase. We are interested in fixed parameter tractable (FPT) algorithms by taking as parameters the dimensions of the blocks and the largest coefficient. For almost combinatorial 4-block n-fold IPs, we first show that there exists some (lambda le g(gamma )) such that for any nonzero kernel element ({textbf{g}}), (lambda {textbf{g}}) can always be decomposed into kernel elements in the same orthant whose (ell _{infty })-norm is bounded by (g(gamma )) (while ({textbf{g}}) itself might not admit such a decomposition), where g is a computable function and (gamma ) is an upper bound on the dimensions of the blocks and the largest coefficient. Based on this, we are able to bound the (ell _{infty })-norm of Graver basis elements by ({mathcal {O}}(g(gamma )n)) and develop an ({mathcal {O}}(g(gamma )n^{3+o(1)}hat{L}^2))-time algorithm (here (hat{L}) denotes the logarithm of the largest absolute value occurring in the input). Additionally, we show that the (ell _{infty })-norm of Graver basis elements is (varOmega (n)). As applications, almost combinatorial 4-block n-fold IPs can be used to model generalizations of classical problems, including scheduling with rejection, bi-criteria scheduling, and a generalized delivery problem. Therefore, our FPT algorithm establishes a general framework to settle these problems.
摘要 本文研究了广义 4 块 n 折 IP 的一个特例,其中 (B_i=B)且 B 的秩最多为 1。这种 IP 被称为近似组合 4 块 n 折 IP,包括广义 n 折 IP 的一个子例。我们感兴趣的是以块的维数和最大系数为参数的固定参数可操作性(FPT)算法。对于几乎是组合型的4块n折叠IP,我们首先证明存在一些 (lambda le g(gamma )) 这样的内核元素:对于任何非零内核元素 ({textbf{g}}) 、 (lambda{textbf{g}})总是可以分解成同一个正交的内核元素,其(ell _{infty }) -norm受(g(gamma ))约束(而({textbf{g}})本身可能不允许这样的分解)、其中,g 是一个可计算的函数,而 (gamma ) 是块的维数和最大系数的上限。在此基础上,我们可以通过 ({mathcal {O}}(g(gamma )n)) 来约束格拉弗基元的 (ell _{infty }) -norm,并开发出一种 ({mathcal {O}}(g(gamma )n^{3+o(1)}hat{L}^2))-时间算法(这里的 (hat{L} 表示输入中出现的最大绝对值的对数)。此外,我们还证明了 Graver 基元的 (ell _{infty }) -norm是 (varOmega (n)) 。作为应用,几乎可以用组合 4 块 n 折 IP 来模拟经典问题的一般化,包括拒绝调度、双标准调度和一般化交付问题。因此,我们的 FPT 算法建立了解决这些问题的通用框架。
{"title":"FPT algorithms for a special block-structured integer program with applications in scheduling","authors":"","doi":"10.1007/s10107-023-02046-z","DOIUrl":"https://doi.org/10.1007/s10107-023-02046-z","url":null,"abstract":"<h3>Abstract</h3> <p>In this paper, a special case of the generalized 4-block <em>n</em>-fold IPs is investigated, where <span> <span>(B_i=B)</span> </span> and <em>B</em> has a rank at most 1. Such IPs, called <em>almost combinatorial 4-block n-fold IPs</em>, include the generalized <em>n</em>-fold IPs as a subcase. We are interested in fixed parameter tractable (FPT) algorithms by taking as parameters the dimensions of the blocks and the largest coefficient. For almost combinatorial 4-block <em>n</em>-fold IPs, we first show that there exists some <span> <span>(lambda le g(gamma ))</span> </span> such that for any nonzero kernel element <span> <span>({textbf{g}})</span> </span>, <span> <span>(lambda {textbf{g}})</span> </span> can always be decomposed into kernel elements in the same orthant whose <span> <span>(ell _{infty })</span> </span>-norm is bounded by <span> <span>(g(gamma ))</span> </span> (while <span> <span>({textbf{g}})</span> </span> itself might not admit such a decomposition), where <em>g</em> is a computable function and <span> <span>(gamma )</span> </span> is an upper bound on the dimensions of the blocks and the largest coefficient. Based on this, we are able to bound the <span> <span>(ell _{infty })</span> </span>-norm of Graver basis elements by <span> <span>({mathcal {O}}(g(gamma )n))</span> </span> and develop an <span> <span>({mathcal {O}}(g(gamma )n^{3+o(1)}hat{L}^2))</span> </span>-time algorithm (here <span> <span>(hat{L})</span> </span> denotes the logarithm of the largest absolute value occurring in the input). Additionally, we show that the <span> <span>(ell _{infty })</span> </span>-norm of Graver basis elements is <span> <span>(varOmega (n))</span> </span>. As applications, almost combinatorial 4-block <em>n</em>-fold IPs can be used to model generalizations of classical problems, including scheduling with rejection, bi-criteria scheduling, and a generalized delivery problem. Therefore, our FPT algorithm establishes a general framework to settle these problems.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"34 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139105010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-04DOI: 10.1007/s10107-023-02041-4
Masoud Ahookhosh, Yurii Nesterov
We introduce a Bi-level OPTimization (BiOPT) framework for minimizing the sum of two convex functions, where one of them is smooth enough. The BiOPT framework offers three levels of freedom: (i) choosing the order p of the proximal term; (ii) designing an inexact pth-order proximal-point method in the upper level; (iii) solving the auxiliary problem with a lower-level non-Euclidean method in the lower level. We here regularize the objective by a ((p+1))th-order proximal term (for arbitrary integer (pge 1)) and then develop the generic inexact high-order proximal-point scheme and its acceleration using the standard estimating sequence technique at the upper level. This follows at the lower level with solving the corresponding pth-order proximal auxiliary problem inexactly either by one iteration of the pth-order tensor method or by a lower-order non-Euclidean composite gradient scheme. Ultimately, it is shown that applying the accelerated inexact pth-order proximal-point method at the upper level and handling the auxiliary problem by the non-Euclidean composite gradient scheme lead to a 2q-order method with the convergence rate ({mathcal {O}}(k^{-(p+1)})) (for (q=lfloor p/2rfloor ) and the iteration counter k), which can result to a superfast method for some specific class of problems.
{"title":"High-order methods beyond the classical complexity bounds: inexact high-order proximal-point methods","authors":"Masoud Ahookhosh, Yurii Nesterov","doi":"10.1007/s10107-023-02041-4","DOIUrl":"https://doi.org/10.1007/s10107-023-02041-4","url":null,"abstract":"<p>We introduce a <i>Bi-level OPTimization</i> (BiOPT) framework for minimizing the sum of two convex functions, where one of them is smooth enough. The BiOPT framework offers three levels of freedom: (i) choosing the order <i>p</i> of the proximal term; (ii) designing an inexact <i>p</i>th-order proximal-point method in the upper level; (iii) solving the auxiliary problem with a lower-level non-Euclidean method in the lower level. We here regularize the objective by a <span>((p+1))</span>th-order proximal term (for arbitrary integer <span>(pge 1)</span>) and then develop the generic inexact high-order proximal-point scheme and its acceleration using the standard estimating sequence technique at the upper level. This follows at the lower level with solving the corresponding <i>p</i>th-order proximal auxiliary problem inexactly either by one iteration of the <i>p</i>th-order tensor method or by a lower-order non-Euclidean composite gradient scheme. Ultimately, it is shown that applying the accelerated inexact <i>p</i>th-order proximal-point method at the upper level and handling the auxiliary problem by the non-Euclidean composite gradient scheme lead to a 2<i>q</i>-order method with the convergence rate <span>({mathcal {O}}(k^{-(p+1)}))</span> (for <span>(q=lfloor p/2rfloor )</span> and the iteration counter <i>k</i>), which can result to a superfast method for some specific class of problems.\u0000</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"65 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139105015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-03DOI: 10.1007/s10107-023-02043-2
Quang Minh Bui, Margarida Carvalho, José Neto
The network pricing problem (NPP) is a bilevel problem, where the leader optimizes its revenue by deciding on the prices of certain arcs in a graph, while expecting the followers (also known as the commodities) to choose a shortest path based on those prices. In this paper, we investigate the complexity of the NPP with respect to two parameters: the number of tolled arcs, and the number of commodities. We devise a simple algorithm showing that if the number of tolled arcs is fixed, then the problem can be solved in polynomial time with respect to the number of commodities. In contrast, even if there is only one commodity, once the number of tolled arcs is not fixed, the problem becomes NP-hard. We characterize this asymmetry in the complexity with a novel property named strong bilevel feasibility. Finally, we describe an algorithm to generate valid inequalities to the NPP based on this property, whose numerical results illustrate its potential for effectively solving the NPP with a high number of commodities.
{"title":"Asymmetry in the complexity of the multi-commodity network pricing problem","authors":"Quang Minh Bui, Margarida Carvalho, José Neto","doi":"10.1007/s10107-023-02043-2","DOIUrl":"https://doi.org/10.1007/s10107-023-02043-2","url":null,"abstract":"<p>The network pricing problem (NPP) is a bilevel problem, where the leader optimizes its revenue by deciding on the prices of certain arcs in a graph, while expecting the followers (also known as the commodities) to choose a shortest path based on those prices. In this paper, we investigate the complexity of the NPP with respect to two parameters: the number of tolled arcs, and the number of commodities. We devise a simple algorithm showing that if the number of tolled arcs is fixed, then the problem can be solved in polynomial time with respect to the number of commodities. In contrast, even if there is only one commodity, once the number of tolled arcs is not fixed, the problem becomes NP-hard. We characterize this asymmetry in the complexity with a novel property named strong bilevel feasibility. Finally, we describe an algorithm to generate valid inequalities to the NPP based on this property, whose numerical results illustrate its potential for effectively solving the NPP with a high number of commodities.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"36 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139094988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-01Epub Date: 2023-03-24DOI: 10.1007/s10107-023-01943-7
Nikita Doikov, Yurii Nesterov
In this paper, we propose a first second-order scheme based on arbitrary non-Euclidean norms, incorporated by Bregman distances. They are introduced directly in the Newton iterate with regularization parameter proportional to the square root of the norm of the current gradient. For the basic scheme, as applied to the composite convex optimization problem, we establish the global convergence rate of the order both in terms of the functional residual and in the norm of subgradients. Our main assumption on the smooth part of the objective is Lipschitz continuity of its Hessian. For uniformly convex functions of degree three, we justify global linear rate, and for strongly convex function we prove the local superlinear rate of convergence. Our approach can be seen as a relaxation of the Cubic Regularization of the Newton method (Nesterov and Polyak in Math Program 108(1):177-205, 2006) for convex minimization problems. This relaxation preserves the convergence properties and global complexities of the Cubic Newton in convex case, while the auxiliary subproblem at each iteration is simpler. We equip our method with adaptive search procedure for choosing the regularization parameter. We propose also an accelerated scheme with convergence rate , where k is the iteration counter.
在本文中,我们提出了一种基于任意非欧几里得规范的一阶二阶方案,并将布雷格曼距离纳入其中。它们直接引入牛顿迭代,正则化参数与当前梯度规范的平方根成正比。对于应用于复合凸优化问题的基本方案,我们确定了在函数残差和子梯度规范方面的全局收敛率为 O(k-2)。我们对目标光滑部分的主要假设是其赫西数的 Lipschitz 连续性。对于三度均匀凸函数,我们证明了全局线性收敛率;对于强凸函数,我们证明了局部超线性收敛率。我们的方法可以看作是牛顿方法立方正则化(Nesterov 和 Polyak 在 Math Program 108(1):177-205, 2006)对凸最小化问题的一种放松。这种松弛保留了凸牛顿法的收敛特性和全局复杂性,而每次迭代的辅助子问题则更为简单。我们的方法采用自适应搜索程序来选择正则化参数。我们还提出了一种收敛速率为 O(k-3)的加速方案,其中 k 是迭代计数器。
{"title":"Gradient regularization of Newton method with Bregman distances.","authors":"Nikita Doikov, Yurii Nesterov","doi":"10.1007/s10107-023-01943-7","DOIUrl":"10.1007/s10107-023-01943-7","url":null,"abstract":"<p><p>In this paper, we propose a first second-order scheme based on arbitrary non-Euclidean norms, incorporated by Bregman distances. They are introduced directly in the Newton iterate with regularization parameter proportional to the square root of the norm of the current gradient. For the basic scheme, as applied to the composite convex optimization problem, we establish the global convergence rate of the order <math><mrow><mi>O</mi><mo>(</mo><msup><mi>k</mi><mrow><mo>-</mo><mn>2</mn></mrow></msup><mo>)</mo></mrow></math> both in terms of the functional residual and in the norm of subgradients. Our main assumption on the smooth part of the objective is Lipschitz continuity of its Hessian. For uniformly convex functions of degree three, we justify global linear rate, and for strongly convex function we prove the local superlinear rate of convergence. Our approach can be seen as a relaxation of the Cubic Regularization of the Newton method (Nesterov and Polyak in Math Program 108(1):177-205, 2006) for convex minimization problems. This relaxation preserves the convergence properties and global complexities of the Cubic Newton in convex case, while the auxiliary subproblem at each iteration is simpler. We equip our method with adaptive search procedure for choosing the regularization parameter. We propose also an accelerated scheme with convergence rate <math><mrow><mi>O</mi><mo>(</mo><msup><mi>k</mi><mrow><mo>-</mo><mn>3</mn></mrow></msup><mo>)</mo></mrow></math>, where <i>k</i> is the iteration counter.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"1 1","pages":"1-25"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10869408/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47300576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-26DOI: 10.1007/s10107-023-02036-1
Volker Krätschmer
We investigate statistical properties of the optimal value of the Sample Average Approximation of stochastic programs, continuing the study (Krätschmer in Nonasymptotic upper estimates for errors of the sample average approximation method to solve risk averse stochastic programs, 2023. Forthcoming in SIAM J. Optim.). Central Limit Theorem type results are derived for the optimal value. As a crucial point the investigations are based on a new type of conditions from the theory of empirical processes which do not rely on pathwise analytical properties of the goal functions. In particular, continuity or convexity in the parameter is not imposed in advance as usual in the literature on the Sample Average Approximation method. It is also shown that the new condition is satisfied if the paths of the goal functions are Hölder continuous so that the main results carry over in this case. Moreover, the main results are applied to goal functions whose paths are piecewise Hölder continuous as e.g. in two stage mixed-integer programs. The main results are shown for classical risk neutral stochastic programs, but we also demonstrate how to apply them to the Sample Average Approximation of risk averse stochastic programs. In this respect we consider stochastic programs expressed in terms of absolute semideviations and divergence risk measures.
我们研究了随机程序的样本平均近似法最优值的统计特性,延续了(Krätschmer 在《解决风险厌恶随机程序的样本平均近似法误差的非渐近上限估计》中的研究,2023.即将发表于 SIAM J. Optim.)。针对最优值推导出了中心极限定理类型的结果。研究的关键点是基于经验过程理论中的新型条件,这些条件并不依赖于目标函数的路径分析特性。特别是,没有像有关样本平均逼近法的文献中通常那样,事先强加参数的连续性或凸性。研究还表明,如果目标函数的路径是荷尔德连续的,那么新条件就会得到满足,因此主要结果在这种情况下也是如此。此外,主要结果还适用于路径为片断荷尔德连续的目标函数,例如两阶段混合整数程序。主要结果针对经典的风险中性随机程序,但我们也演示了如何将这些结果应用于风险规避随机程序的抽样平均逼近。在这方面,我们考虑了以绝对半偏差和发散风险度量表示的随机程序。
{"title":"First order asymptotics of the sample average approximation method to solve risk averse stochastic programs","authors":"Volker Krätschmer","doi":"10.1007/s10107-023-02036-1","DOIUrl":"https://doi.org/10.1007/s10107-023-02036-1","url":null,"abstract":"<p>We investigate statistical properties of the optimal value of the Sample Average Approximation of stochastic programs, continuing the study (Krätschmer in Nonasymptotic upper estimates for errors of the sample average approximation method to solve risk averse stochastic programs, 2023. Forthcoming in SIAM J. Optim.). Central Limit Theorem type results are derived for the optimal value. As a crucial point the investigations are based on a new type of conditions from the theory of empirical processes which do not rely on pathwise analytical properties of the goal functions. In particular, continuity or convexity in the parameter is not imposed in advance as usual in the literature on the Sample Average Approximation method. It is also shown that the new condition is satisfied if the paths of the goal functions are Hölder continuous so that the main results carry over in this case. Moreover, the main results are applied to goal functions whose paths are piecewise Hölder continuous as e.g. in two stage mixed-integer programs. The main results are shown for classical risk neutral stochastic programs, but we also demonstrate how to apply them to the Sample Average Approximation of risk averse stochastic programs. In this respect we consider stochastic programs expressed in terms of absolute semideviations and divergence risk measures.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"15 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2023-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139052819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-26DOI: 10.1007/s10107-023-02037-0
Moslem Zamani, Hadi Abbaszadehpeivasti, Etienne de Klerk
Recently, semidefinite programming performance estimation has been employed as a strong tool for the worst-case performance analysis of first order methods. In this paper, we derive new non-ergodic convergence rates for the alternating direction method of multipliers (ADMM) by using performance estimation. We give some examples which show the exactness of the given bounds. We also study the linear and R-linear convergence of ADMM in terms of dual objective. We establish that ADMM enjoys a global linear convergence rate if and only if the dual objective satisfies the Polyak–Łojasiewicz (PŁ) inequality in the presence of strong convexity. In addition, we give an explicit formula for the linear convergence rate factor. Moreover, we study the R-linear convergence of ADMM under two scenarios.
最近,半定量编程性能估计被用作一阶方法最坏情况性能分析的有力工具。在本文中,我们利用性能估计推导出了交替方向乘法(ADMM)新的非啮合收敛率。我们举例说明了所给边界的精确性。我们还研究了 ADMM 在双重目标方面的线性和 R 线性收敛性。我们确定,当且仅当对偶目标满足强凸性条件下的 Polyak-Łojasiewicz (PŁ) 不等式时,ADMM 才享有全局线性收敛率。此外,我们还给出了线性收敛率因子的明确公式。此外,我们还研究了两种情况下 ADMM 的 R 线性收敛。
{"title":"The exact worst-case convergence rate of the alternating direction method of multipliers","authors":"Moslem Zamani, Hadi Abbaszadehpeivasti, Etienne de Klerk","doi":"10.1007/s10107-023-02037-0","DOIUrl":"https://doi.org/10.1007/s10107-023-02037-0","url":null,"abstract":"<p>Recently, semidefinite programming performance estimation has been employed as a strong tool for the worst-case performance analysis of first order methods. In this paper, we derive new non-ergodic convergence rates for the alternating direction method of multipliers (ADMM) by using performance estimation. We give some examples which show the exactness of the given bounds. We also study the linear and R-linear convergence of ADMM in terms of dual objective. We establish that ADMM enjoys a global linear convergence rate if and only if the dual objective satisfies the Polyak–Łojasiewicz (PŁ) inequality in the presence of strong convexity. In addition, we give an explicit formula for the linear convergence rate factor. Moreover, we study the R-linear convergence of ADMM under two scenarios.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"15 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2023-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139052492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-21DOI: 10.1007/s10107-023-02030-7
Abstract
The truncated singular value decomposition (SVD), also known as the best low-rank matrix approximation with minimum error measured by a unitarily invariant norm, has been applied to many domains such as biology, healthcare, among others, where high-dimensional datasets are prevalent. To extract interpretable information from the high-dimensional data, sparse truncated SVD (SSVD) has been used to select a handful of rows and columns of the original matrix along with the best low-rank approximation. Different from the literature on SSVD focusing on the top singular value or compromising the sparsity for the seek of computational efficiency, this paper presents a novel SSVD formulation that can select the best submatrix precisely up to a given size to maximize its truncated Ky Fan norm. The fact that the proposed SSVD problem is NP-hard motivates us to study effective algorithms with provable performance guarantees. To do so, we first reformulate SSVD as a mixed-integer semidefinite program, which can be solved exactly for small- or medium-sized instances within a branch-and-cut algorithm framework with closed-form cuts and is extremely useful for evaluating the solution quality of approximation algorithms. We next develop three selection algorithms based on different selection criteria and two searching algorithms, greedy and local search. We prove the approximation ratios for all the approximation algorithms and show that all the ratios are tight when the number of rows or columns of the selected submatrix is no larger than half of the data matrix, i.e., our derived approximation ratios are unimprovable. Our numerical study demonstrates the high solution quality and computational efficiency of the proposed algorithms. Finally, all our analysis can be extended to row-sparse PCA.
{"title":"Beyond symmetry: best submatrix selection for the sparse truncated SVD","authors":"","doi":"10.1007/s10107-023-02030-7","DOIUrl":"https://doi.org/10.1007/s10107-023-02030-7","url":null,"abstract":"<h3>Abstract</h3> <p>The truncated singular value decomposition (SVD), also known as the best low-rank matrix approximation with minimum error measured by a unitarily invariant norm, has been applied to many domains such as biology, healthcare, among others, where high-dimensional datasets are prevalent. To extract interpretable information from the high-dimensional data, sparse truncated SVD (SSVD) has been used to select a handful of rows and columns of the original matrix along with the best low-rank approximation. Different from the literature on SSVD focusing on the top singular value or compromising the sparsity for the seek of computational efficiency, this paper presents a novel SSVD formulation that can select the best submatrix precisely up to a given size to maximize its truncated Ky Fan norm. The fact that the proposed SSVD problem is NP-hard motivates us to study effective algorithms with provable performance guarantees. To do so, we first reformulate SSVD as a mixed-integer semidefinite program, which can be solved exactly for small- or medium-sized instances within a branch-and-cut algorithm framework with closed-form cuts and is extremely useful for evaluating the solution quality of approximation algorithms. We next develop three selection algorithms based on different selection criteria and two searching algorithms, greedy and local search. We prove the approximation ratios for all the approximation algorithms and show that all the ratios are tight when the number of rows or columns of the selected submatrix is no larger than half of the data matrix, i.e., our derived approximation ratios are unimprovable. Our numerical study demonstrates the high solution quality and computational efficiency of the proposed algorithms. Finally, all our analysis can be extended to row-sparse PCA.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"13 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2023-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138825723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-18DOI: 10.1007/s10107-023-02034-3
Martin Drees
A clutter is a family of sets, called members, such that no member contains another. It is called intersecting if every two members intersect, but not all members have a common element. Dense clutters additionally do not have a fractional packing of value 2. We are looking at certain substructures of clutters, namely minors and restrictions. For a family of clutters we introduce a general sufficient condition such that for every clutter we can decide whether the clutter has a restriction in that set in polynomial time. It is known that the sets of intersecting and dense clutters satisfy this condition. For intersecting clutters we generalize the statement to k-wise intersecting clutters using a much simpler proof. We also give a simplified proof that a dense clutter with no proper dense minor is either a delta or the blocker of an extended odd hole. This simplification reduces the running time of the algorithm for finding a delta or the blocker of an extended odd hole minor from previously ({mathscr {O}}(n^4)) to ({mathscr {O}}(n^3)) filter oracle calls.
{"title":"Intersecting and dense restrictions of clutters in polynomial time","authors":"Martin Drees","doi":"10.1007/s10107-023-02034-3","DOIUrl":"https://doi.org/10.1007/s10107-023-02034-3","url":null,"abstract":"<p>A clutter is a family of sets, called members, such that no member contains another. It is called intersecting if every two members intersect, but not all members have a common element. Dense clutters additionally do not have a fractional packing of value 2. We are looking at certain substructures of clutters, namely minors and restrictions. For a family of clutters we introduce a general sufficient condition such that for every clutter we can decide whether the clutter has a restriction in that set in polynomial time. It is known that the sets of intersecting and dense clutters satisfy this condition. For intersecting clutters we generalize the statement to <i>k</i>-wise intersecting clutters using a much simpler proof. We also give a simplified proof that a dense clutter with no proper dense minor is either a delta or the blocker of an extended odd hole. This simplification reduces the running time of the algorithm for finding a delta or the blocker of an extended odd hole minor from previously <span>({mathscr {O}}(n^4))</span> to <span>({mathscr {O}}(n^3))</span> filter oracle calls.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"20 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2023-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138715571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-14DOI: 10.1007/s10107-023-02038-z
Majid Farhadi, Swati Gupta, Shengding Sun, Prasad Tetali, Michael C. Wigal
The minimum linear ordering problem (MLOP) generalizes well-known combinatorial optimization problems such as minimum linear arrangement and minimum sum set cover. MLOP seeks to minimize an aggregated cost (f(cdot )) due to an ordering (sigma ) of the items (say [n]), i.e., (min _{sigma } sum _{iin [n]} f(E_{i,sigma })), where (E_{i,sigma }) is the set of items mapped by (sigma ) to indices [i]. Despite an extensive literature on MLOP variants and approximations for these, it was unclear whether the graphic matroid MLOP was NP-hard. We settle this question through non-trivial reductions from mininimum latency vertex cover and minimum sum vertex cover problems. We further propose a new combinatorial algorithm for approximating monotone submodular MLOP, using the theory of principal partitions. This is in contrast to the rounding algorithm by Iwata et al. (in: APPROX, 2012), using Lovász extension of submodular functions. We show a ((2-frac{1+ell _{f}}{1+|E|}))-approximation for monotone submodular MLOP where (ell _{f}=frac{f(E)}{max _{xin E}f({x})}) satisfies (1 le ell _f le |E|). Our theory provides new approximation bounds for special cases of the problem, in particular a ((2-frac{1+r(E)}{1+|E|}))-approximation for the matroid MLOP, where (f = r) is the rank function of a matroid. We further show that minimum latency vertex cover is (frac{4}{3})-approximable, by which we also lower bound the integrality gap of its natural LP relaxation, which might be of independent interest.
{"title":"Hardness and approximation of submodular minimum linear ordering problems","authors":"Majid Farhadi, Swati Gupta, Shengding Sun, Prasad Tetali, Michael C. Wigal","doi":"10.1007/s10107-023-02038-z","DOIUrl":"https://doi.org/10.1007/s10107-023-02038-z","url":null,"abstract":"<p>The minimum linear ordering problem (MLOP) generalizes well-known combinatorial optimization problems such as minimum linear arrangement and minimum sum set cover. MLOP seeks to minimize an aggregated cost <span>(f(cdot ))</span> due to an ordering <span>(sigma )</span> of the items (say [<i>n</i>]), i.e., <span>(min _{sigma } sum _{iin [n]} f(E_{i,sigma }))</span>, where <span>(E_{i,sigma })</span> is the set of items mapped by <span>(sigma )</span> to indices [<i>i</i>]. Despite an extensive literature on MLOP variants and approximations for these, it was unclear whether the graphic matroid MLOP was NP-hard. We settle this question through non-trivial reductions from mininimum latency vertex cover and minimum sum vertex cover problems. We further propose a new combinatorial algorithm for approximating monotone submodular MLOP, using the theory of principal partitions. This is in contrast to the rounding algorithm by Iwata et al. (in: APPROX, 2012), using Lovász extension of submodular functions. We show a <span>((2-frac{1+ell _{f}}{1+|E|}))</span>-approximation for monotone submodular MLOP where <span>(ell _{f}=frac{f(E)}{max _{xin E}f({x})})</span> satisfies <span>(1 le ell _f le |E|)</span>. Our theory provides new approximation bounds for special cases of the problem, in particular a <span>((2-frac{1+r(E)}{1+|E|}))</span>-approximation for the matroid MLOP, where <span>(f = r)</span> is the rank function of a matroid. We further show that minimum latency vertex cover is <span>(frac{4}{3})</span>-approximable, by which we also lower bound the integrality gap of its natural LP relaxation, which might be of independent interest.</p>","PeriodicalId":18297,"journal":{"name":"Mathematical Programming","volume":"200 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2023-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138690214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}