首页 > 最新文献

SIAM Journal on Optimization最新文献

英文 中文
Exact Quantization of Multistage Stochastic Linear Problems 多阶段随机线性问题的精确量化
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-02-05 DOI: 10.1137/22m1508005
Maël Forcier, Stéphane Gaubert, Vincent Leclère
SIAM Journal on Optimization, Volume 34, Issue 1, Page 533-562, March 2024.
Abstract. We show that the multistage stochastic linear problem (MSLP) with an arbitrary cost distribution is equivalent to an MSLP on a finite scenario tree. We establish this exact quantization result by analyzing the polyhedral structure of MSLPs. In particular, we show that the expected cost-to-go functions are polyhedral and affine on the cells of a chamber complex, which is independent of the cost distribution. This leads to new complexity results, showing that MSLP becomes polynomial when certain parameters are fixed.
SIAM 优化期刊》,第 34 卷,第 1 期,第 533-562 页,2024 年 3 月。 摘要我们证明了具有任意成本分布的多阶段随机线性问题(MSLP)等价于有限情景树上的 MSLP。我们通过分析 MSLP 的多面体结构建立了这一精确量化结果。特别是,我们证明了预期成本-去向函数是多面体的,并且在室复数的单元上是仿射的,这与成本分布无关。这带来了新的复杂性结果,表明当某些参数固定时,MSLP 会变成多项式。
{"title":"Exact Quantization of Multistage Stochastic Linear Problems","authors":"Maël Forcier, Stéphane Gaubert, Vincent Leclère","doi":"10.1137/22m1508005","DOIUrl":"https://doi.org/10.1137/22m1508005","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 533-562, March 2024. <br/> Abstract. We show that the multistage stochastic linear problem (MSLP) with an arbitrary cost distribution is equivalent to an MSLP on a finite scenario tree. We establish this exact quantization result by analyzing the polyhedral structure of MSLPs. In particular, we show that the expected cost-to-go functions are polyhedral and affine on the cells of a chamber complex, which is independent of the cost distribution. This leads to new complexity results, showing that MSLP becomes polynomial when certain parameters are fixed.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"69 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139765297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Algorithms for Finding a D-Stationary Point of a Class of Structured Nonsmooth DC Minimization 寻找一类结构非光滑直流最小化的 D-静态点的混合算法
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-02-01 DOI: 10.1137/21m1457709
Zhe Sun, Lei Wu
SIAM Journal on Optimization, Volume 34, Issue 1, Page 485-506, March 2024.
Abstract. In this paper, we consider a class of structured nonsmooth difference-of-convex (DC) minimization in which the first convex component is the sum of a smooth and a nonsmooth function, while the second convex component is the supremum of finitely many convex smooth functions. The existing methods for this problem usually have weak convergence guarantees or need to solve lots of subproblems per iteration. Due to this, we propose hybrid algorithms for solving this problem in which we first compute approximate critical points and then check whether these points are approximate D-stationary points. Under suitable conditions, we prove that there exists a subsequence of iterates of which every accumulation point is a D-stationary point. Some preliminary numerical experiments are conducted to demonstrate the efficiency of the proposed algorithms.
SIAM 优化期刊》,第 34 卷第 1 期,第 485-506 页,2024 年 3 月。 摘要本文考虑了一类结构化非光滑凸差(DC)最小化问题,其中第一个凸分量是光滑函数与非光滑函数之和,而第二个凸分量是有限多个凸光滑函数的上集。针对这一问题的现有方法通常收敛保证较弱,或者每次迭代需要解决大量子问题。因此,我们提出了解决该问题的混合算法,即首先计算近似临界点,然后检查这些点是否为近似 D-station 点。在合适的条件下,我们证明存在一个迭代的子序列,其中的每个累积点都是 D 静止点。我们还进行了一些初步的数值实验,以证明所提算法的效率。
{"title":"Hybrid Algorithms for Finding a D-Stationary Point of a Class of Structured Nonsmooth DC Minimization","authors":"Zhe Sun, Lei Wu","doi":"10.1137/21m1457709","DOIUrl":"https://doi.org/10.1137/21m1457709","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 485-506, March 2024. <br/> Abstract. In this paper, we consider a class of structured nonsmooth difference-of-convex (DC) minimization in which the first convex component is the sum of a smooth and a nonsmooth function, while the second convex component is the supremum of finitely many convex smooth functions. The existing methods for this problem usually have weak convergence guarantees or need to solve lots of subproblems per iteration. Due to this, we propose hybrid algorithms for solving this problem in which we first compute approximate critical points and then check whether these points are approximate D-stationary points. Under suitable conditions, we prove that there exists a subsequence of iterates of which every accumulation point is a D-stationary point. Some preliminary numerical experiments are conducted to demonstrate the efficiency of the proposed algorithms.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"10 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139669745","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shortest Paths in Graphs of Convex Sets 凸集合图中的最短路径
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-02-01 DOI: 10.1137/22m1523790
Tobia Marcucci, Jack Umenberger, Pablo Parrilo, Russ Tedrake
SIAM Journal on Optimization, Volume 34, Issue 1, Page 507-532, March 2024.
Abstract. Given a graph, the shortest-path problem requires finding a sequence of edges with minimum cumulative length that connects a source vertex to a target vertex. We consider a variant of this classical problem in which the position of each vertex in the graph is a continuous decision variable constrained in a convex set, and the length of an edge is a convex function of the position of its endpoints. Problems of this form arise naturally in many areas, from motion planning of autonomous vehicles to optimal control of hybrid systems. The price for such a wide applicability is the complexity of this problem, which is easily seen to be NP-hard. Our main contribution is a strong and lightweight mixed-integer convex formulation based on perspective operators, that makes it possible to efficiently find globally optimal paths in large graphs and in high-dimensional spaces.
SIAM 优化期刊》,第 34 卷第 1 期,第 507-532 页,2024 年 3 月。 摘要给定一个图,最短路径问题要求找到一个连接源顶点和目标顶点的累积长度最小的边序列。我们考虑了这一经典问题的一个变体,其中图中每个顶点的位置是一个连续的决策变量,受限于一个凸集,而边的长度是其端点位置的凸函数。从自动驾驶汽车的运动规划到混合动力系统的优化控制,这种形式的问题自然出现在许多领域。这种广泛适用性的代价是这一问题的复杂性,很容易被认为是 NP 难。我们的主要贡献是基于透视算子的强大而轻便的混合整数凸表述,它使得在大型图和高维空间中高效地找到全局最优路径成为可能。
{"title":"Shortest Paths in Graphs of Convex Sets","authors":"Tobia Marcucci, Jack Umenberger, Pablo Parrilo, Russ Tedrake","doi":"10.1137/22m1523790","DOIUrl":"https://doi.org/10.1137/22m1523790","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 507-532, March 2024. <br/> Abstract. Given a graph, the shortest-path problem requires finding a sequence of edges with minimum cumulative length that connects a source vertex to a target vertex. We consider a variant of this classical problem in which the position of each vertex in the graph is a continuous decision variable constrained in a convex set, and the length of an edge is a convex function of the position of its endpoints. Problems of this form arise naturally in many areas, from motion planning of autonomous vehicles to optimal control of hybrid systems. The price for such a wide applicability is the complexity of this problem, which is easily seen to be NP-hard. Our main contribution is a strong and lightweight mixed-integer convex formulation based on perspective operators, that makes it possible to efficiently find globally optimal paths in large graphs and in high-dimensional spaces.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"38 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139669374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Infeasibility Detection with Primal-Dual Hybrid Gradient for Large-Scale Linear Programming 大规模线性规划的原点-双混合梯度不可行性检测
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-31 DOI: 10.1137/22m1510467
David Applegate, Mateo Díaz, Haihao Lu, Miles Lubin
SIAM Journal on Optimization, Volume 34, Issue 1, Page 459-484, March 2024.
Abstract. We study the problem of detecting infeasibility of large-scale linear programming problems using the primal-dual hybrid gradient (PDHG) method of Chambolle and Pock [J. Math. Imaging Vision, 40 (2011), pp. 120–145]. The literature on PDHG has focused chiefly on problems with at least one optimal solution. We show that when the problem is infeasible or unbounded, the iterates diverge at a controlled rate toward a well-defined ray. In turn, the direction of such a ray recovers infeasibility certificates. Based on this fact, we propose a simple way to extract approximate infeasibility certificates from the iterates of PDHG. We study three sequences that converge to certificates: the difference of iterates, the normalized iterates, and the normalized average. All of them are easy to compute and suitable for large-scale problems. We show that the normalized iterates and normalized averages achieve a convergence rate of [math]. This rate is general and applies to any fixed-point iteration of a nonexpansive operator. Thus, it is a result of independent interest that goes well beyond our setting. Finally, we show that, under nondegeneracy assumptions, the iterates of PDHG identify the active set of an auxiliary feasible problem in finite time, which ensures that the difference of iterates exhibits eventual linear convergence. These results provide a theoretical justification for infeasibility detection in the newly developed linear programming solver PDLP.
SIAM 优化期刊》,第 34 卷第 1 期,第 459-484 页,2024 年 3 月。 摘要我们使用 Chambolle 和 Pock [J. Math. Imaging Vision, 40 (2011), pp.关于 PDHG 的文献主要集中在至少有一个最优解的问题上。我们的研究表明,当问题不可行或无边界时,迭代会以可控的速度向一条定义明确的射线发散。反过来,这种射线的方向也能恢复不可行性证明。基于这一事实,我们提出了一种从 PDHG 迭代中提取近似不可行性证明的简单方法。我们研究了收敛到证书的三个序列:迭代差、归一化迭代和归一化平均。它们都易于计算,适用于大规模问题。我们证明,归一化迭代和归一化平均达到了 [math] 的收敛速率。这个收敛率是通用的,适用于非展开算子的任何定点迭代。因此,它是一个独立的结果,远远超出了我们的设定。最后,我们证明,在非孤立性假设下,PDHG 的迭代在有限时间内确定了辅助可行问题的活动集,这确保了迭代差最终呈现线性收敛。这些结果为新开发的线性规划求解器 PDLP 中的不可行性检测提供了理论依据。
{"title":"Infeasibility Detection with Primal-Dual Hybrid Gradient for Large-Scale Linear Programming","authors":"David Applegate, Mateo Díaz, Haihao Lu, Miles Lubin","doi":"10.1137/22m1510467","DOIUrl":"https://doi.org/10.1137/22m1510467","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 459-484, March 2024. <br/> Abstract. We study the problem of detecting infeasibility of large-scale linear programming problems using the primal-dual hybrid gradient (PDHG) method of Chambolle and Pock [J. Math. Imaging Vision, 40 (2011), pp. 120–145]. The literature on PDHG has focused chiefly on problems with at least one optimal solution. We show that when the problem is infeasible or unbounded, the iterates diverge at a controlled rate toward a well-defined ray. In turn, the direction of such a ray recovers infeasibility certificates. Based on this fact, we propose a simple way to extract approximate infeasibility certificates from the iterates of PDHG. We study three sequences that converge to certificates: the difference of iterates, the normalized iterates, and the normalized average. All of them are easy to compute and suitable for large-scale problems. We show that the normalized iterates and normalized averages achieve a convergence rate of [math]. This rate is general and applies to any fixed-point iteration of a nonexpansive operator. Thus, it is a result of independent interest that goes well beyond our setting. Finally, we show that, under nondegeneracy assumptions, the iterates of PDHG identify the active set of an auxiliary feasible problem in finite time, which ensures that the difference of iterates exhibits eventual linear convergence. These results provide a theoretical justification for infeasibility detection in the newly developed linear programming solver PDLP.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"26 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139657357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributionally Favorable Optimization: A Framework for Data-Driven Decision-Making with Endogenous Outliers 有利于分布的优化:内生异常值数据驱动决策框架
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-29 DOI: 10.1137/22m1528094
Nan Jiang, Weijun Xie
SIAM Journal on Optimization, Volume 34, Issue 1, Page 419-458, March 2024.
Abstract. A typical data-driven stochastic program seeks the best decision that minimizes the sum of a deterministic cost function and an expected recourse function under a given distribution. Recently, much success has been witnessed in the development of distributionally robust optimization (DRO), which considers the worst-case expected recourse function under the least favorable probability distribution from a distributional family. However, in the presence of endogenous outliers such that their corresponding recourse function values are very large or even infinite, the commonly used DRO framework alone tends to overemphasize these endogenous outliers and cause undesirable or even infeasible decisions. On the contrary, distributionally favorable optimization (DFO), concerning the best-case expected recourse function under the most favorable distribution from the distributional family, can serve as a proper measure of the stochastic recourse function and mitigate the effect of endogenous outliers. We show that DFO recovers many robust statistics, suggesting that the DFO framework might be appropriate for the stochastic recourse function in the presence of endogenous outliers. A notion of decision outlier robustness is proposed for selecting a DFO framework for data-driven optimization with outliers. We also provide a unified way to integrate DRO with DFO, where DRO addresses the out-of-sample performance, and DFO properly handles the stochastic recourse function under endogenous outliers. We further extend the proposed DFO framework to solve two-stage stochastic programs without relatively complete recourse. The numerical study demonstrates that the framework is promising.
SIAM 优化期刊》,第 34 卷,第 1 期,第 419-458 页,2024 年 3 月。 摘要。典型的数据驱动随机程序寻求在给定分布条件下使确定性成本函数与期望求助函数之和最小化的最佳决策。最近,分布稳健优化(DRO)的发展取得了巨大成功,它考虑了分布族中最不利概率分布下的最坏情况预期求助函数。然而,在存在内生异常值的情况下,其相应的求助函数值非常大,甚至是无限大,仅靠常用的分布鲁棒优化框架往往会过度强调这些内生异常值,从而导致不理想甚至不可行的决策。相反,分布有利优化(DFO)涉及分布族中最有利分布下的最佳预期求助函数,可以作为随机求助函数的适当度量,并减轻内生异常值的影响。我们的研究表明,DFO 恢复了许多稳健的统计数据,这表明 DFO 框架可能适用于存在内生异常值的随机求助函数。我们提出了决策离群稳健性的概念,以便为有离群值的数据驱动优化选择 DFO 框架。我们还提供了一种整合 DRO 和 DFO 的统一方法,其中 DRO 解决样本外性能问题,DFO 妥善处理内生异常值下的随机求助函数。我们进一步扩展了所提出的 DFO 框架,以求解没有相对完全追索权的两阶段随机程序。数值研究表明,该框架前景广阔。
{"title":"Distributionally Favorable Optimization: A Framework for Data-Driven Decision-Making with Endogenous Outliers","authors":"Nan Jiang, Weijun Xie","doi":"10.1137/22m1528094","DOIUrl":"https://doi.org/10.1137/22m1528094","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 419-458, March 2024. <br/> Abstract. A typical data-driven stochastic program seeks the best decision that minimizes the sum of a deterministic cost function and an expected recourse function under a given distribution. Recently, much success has been witnessed in the development of distributionally robust optimization (DRO), which considers the worst-case expected recourse function under the least favorable probability distribution from a distributional family. However, in the presence of endogenous outliers such that their corresponding recourse function values are very large or even infinite, the commonly used DRO framework alone tends to overemphasize these endogenous outliers and cause undesirable or even infeasible decisions. On the contrary, distributionally favorable optimization (DFO), concerning the best-case expected recourse function under the most favorable distribution from the distributional family, can serve as a proper measure of the stochastic recourse function and mitigate the effect of endogenous outliers. We show that DFO recovers many robust statistics, suggesting that the DFO framework might be appropriate for the stochastic recourse function in the presence of endogenous outliers. A notion of decision outlier robustness is proposed for selecting a DFO framework for data-driven optimization with outliers. We also provide a unified way to integrate DRO with DFO, where DRO addresses the out-of-sample performance, and DFO properly handles the stochastic recourse function under endogenous outliers. We further extend the proposed DFO framework to solve two-stage stochastic programs without relatively complete recourse. The numerical study demonstrates that the framework is promising.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"46 2 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139585526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian Stochastic Gradient Descent for Stochastic Optimization with Streaming Input Data 利用流输入数据进行随机优化的贝叶斯随机梯度下降法
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-25 DOI: 10.1137/22m1478951
Tianyi Liu, Yifan Lin, Enlu Zhou
SIAM Journal on Optimization, Volume 34, Issue 1, Page 389-418, March 2024.
Abstract. We consider stochastic optimization under distributional uncertainty, where the unknown distributional parameter is estimated from streaming data that arrive sequentially over time. Moreover, data may depend on the decision at the time when they are generated. For both decision-independent and decision-dependent uncertainties, we propose an approach to jointly estimate the distributional parameter via Bayesian posterior distribution and update the decision by applying stochastic gradient descent (SGD) on the Bayesian average of the objective function. Our approach converges asymptotically over time and achieves the convergence rates of classical SGD in the decision-independent case. We demonstrate the empirical performance of our approach on both synthetic test problems and a classical newsvendor problem.
SIAM 优化期刊》,第 34 卷,第 1 期,第 389-418 页,2024 年 3 月。 摘要。我们考虑分布不确定性下的随机优化,其中未知的分布参数是从随时间顺序到达的流数据中估计出来的。此外,数据在生成时可能取决于决策。对于与决策无关和与决策有关的不确定性,我们提出了一种通过贝叶斯后验分布联合估计分布参数的方法,并通过对目标函数的贝叶斯平均值应用随机梯度下降(SGD)来更新决策。我们的方法随时间渐进收敛,并在决策无关的情况下达到经典 SGD 的收敛率。我们在合成测试问题和经典新闻供应商问题上演示了我们方法的经验性能。
{"title":"Bayesian Stochastic Gradient Descent for Stochastic Optimization with Streaming Input Data","authors":"Tianyi Liu, Yifan Lin, Enlu Zhou","doi":"10.1137/22m1478951","DOIUrl":"https://doi.org/10.1137/22m1478951","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 389-418, March 2024. <br/> Abstract. We consider stochastic optimization under distributional uncertainty, where the unknown distributional parameter is estimated from streaming data that arrive sequentially over time. Moreover, data may depend on the decision at the time when they are generated. For both decision-independent and decision-dependent uncertainties, we propose an approach to jointly estimate the distributional parameter via Bayesian posterior distribution and update the decision by applying stochastic gradient descent (SGD) on the Bayesian average of the objective function. Our approach converges asymptotically over time and achieves the convergence rates of classical SGD in the decision-independent case. We demonstrate the empirical performance of our approach on both synthetic test problems and a classical newsvendor problem.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"214 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139551719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Descent Properties of an Anderson Accelerated Gradient Method with Restarting 带重启的安德森加速梯度法的下降特性
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-19 DOI: 10.1137/22m151460x
Wenqing Ouyang, Yang Liu, Andre Milzarek
SIAM Journal on Optimization, Volume 34, Issue 1, Page 336-365, March 2024.
Abstract. Anderson acceleration ([math]) is a popular acceleration technique to enhance the convergence of fixed-point schemes. The analysis of [math] approaches often focuses on the convergence behavior of a corresponding fixed-point residual, while the behavior of the underlying objective function values along the accelerated iterates is currently not well understood. In this paper, we investigate local properties of [math] with restarting applied to a basic gradient scheme ([math]) in terms of function values. Specifically, we show that [math] is a local descent method and that it can decrease the objective function at a rate no slower than the gradient method up to higher-order error terms. These new results theoretically support the good numerical performance of [math] when heuristic descent conditions are used for globalization and they provide a novel perspective on the convergence analysis of [math] that is more amenable to nonconvex optimization problems. Numerical experiments are conducted to illustrate our theoretical findings.
SIAM 优化期刊》,第 34 卷,第 1 期,第 336-365 页,2024 年 3 月。 摘要安德森加速([math])是一种流行的加速技术,用于提高定点方案的收敛性。对[math]方法的分析通常集中在相应定点残差的收敛行为上,而对加速迭代过程中基本目标函数值的行为目前还不甚了解。在本文中,我们从函数值的角度研究了应用于基本梯度方案([math])的[math]重启的局部特性。具体来说,我们证明了 [math] 是一种局部下降方法,它能以不慢于梯度方法的速度减少目标函数,直至高阶误差项。这些新结果从理论上支持了[math]在使用启发式下降条件进行全局化时的良好数值性能,并为[math]的收敛分析提供了一个新的视角,更适合于非凸优化问题。我们进行了数值实验来说明我们的理论发现。
{"title":"Descent Properties of an Anderson Accelerated Gradient Method with Restarting","authors":"Wenqing Ouyang, Yang Liu, Andre Milzarek","doi":"10.1137/22m151460x","DOIUrl":"https://doi.org/10.1137/22m151460x","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 336-365, March 2024. <br/> Abstract. Anderson acceleration ([math]) is a popular acceleration technique to enhance the convergence of fixed-point schemes. The analysis of [math] approaches often focuses on the convergence behavior of a corresponding fixed-point residual, while the behavior of the underlying objective function values along the accelerated iterates is currently not well understood. In this paper, we investigate local properties of [math] with restarting applied to a basic gradient scheme ([math]) in terms of function values. Specifically, we show that [math] is a local descent method and that it can decrease the objective function at a rate no slower than the gradient method up to higher-order error terms. These new results theoretically support the good numerical performance of [math] when heuristic descent conditions are used for globalization and they provide a novel perspective on the convergence analysis of [math] that is more amenable to nonconvex optimization problems. Numerical experiments are conducted to illustrate our theoretical findings.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"3 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139495212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Basic Convex Analysis in Metric Spaces with Bounded Curvature 有界曲率公元空间中的基本凸分析
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-19 DOI: 10.1137/23m1551389
Adrian S. Lewis, Genaro López-Acedo, Adriana Nicolae
SIAM Journal on Optimization, Volume 34, Issue 1, Page 366-388, March 2024.
Abstract. Differentiable structure ensures that many of the basics of classical convex analysis extend naturally from Euclidean space to Riemannian manifolds. Without such structure, however, extensions are more challenging. Nonetheless, in Alexandrov spaces with curvature bounded above (but possibly positive), we develop several basic building blocks. We define subgradients via projection and the normal cone, prove their existence, and relate them to the classical affine minorant property. Then, in what amounts to a simple calculus or duality result, we develop a necessary optimality condition for minimizing the sum of two convex functions.
SIAM 优化期刊》,第 34 卷,第 1 期,第 366-388 页,2024 年 3 月。 摘要可微分结构确保经典凸分析的许多基本原理从欧几里得空间自然扩展到黎曼流形。然而,如果没有这样的结构,扩展就更具挑战性。尽管如此,在曲率有上界(但可能是正)的亚历山德罗夫空间中,我们开发了几个基本的构建模块。我们通过投影和法锥定义了子梯度,证明了它们的存在,并将它们与经典仿射微分性质联系起来。然后,在相当于一个简单的微积分或对偶性结果中,我们提出了最小化两个凸函数之和的必要最优条件。
{"title":"Basic Convex Analysis in Metric Spaces with Bounded Curvature","authors":"Adrian S. Lewis, Genaro López-Acedo, Adriana Nicolae","doi":"10.1137/23m1551389","DOIUrl":"https://doi.org/10.1137/23m1551389","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 366-388, March 2024. <br/> Abstract. Differentiable structure ensures that many of the basics of classical convex analysis extend naturally from Euclidean space to Riemannian manifolds. Without such structure, however, extensions are more challenging. Nonetheless, in Alexandrov spaces with curvature bounded above (but possibly positive), we develop several basic building blocks. We define subgradients via projection and the normal cone, prove their existence, and relate them to the classical affine minorant property. Then, in what amounts to a simple calculus or duality result, we develop a necessary optimality condition for minimizing the sum of two convex functions.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"1 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139495059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Decomposition Algorithm for Two-Stage Stochastic Programs with Nonconvex Recourse Functions 具有非凸求助函数的两阶段随机程序的分解算法
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-19 DOI: 10.1137/22m1488533
Hanyang Li, Ying Cui
SIAM Journal on Optimization, Volume 34, Issue 1, Page 306-335, March 2024.
Abstract. In this paper, we have studied a decomposition method for solving a class of nonconvex two-stage stochastic programs, where both the objective and constraints of the second-stage problem are nonlinearly parameterized by the first-stage variables. Due to the failure of the Clarke regularity of the resulting nonconvex recourse function, classical decomposition approaches such as Benders decomposition and (augmented) Lagrangian-based algorithms cannot be directly generalized to solve such models. By exploring an implicitly convex-concave structure of the recourse function, we introduce a novel decomposition framework based on the so-called partial Moreau envelope. The algorithm successively generates strongly convex quadratic approximations of the recourse function based on the solutions of the second-stage convex subproblems and adds them to the first-stage master problem. Convergence has been established for both a fixed number of scenarios and a sequential internal sampling strategy. Numerical experiments are conducted to demonstrate the effectiveness of the proposed algorithm.
SIAM 优化期刊》,第 34 卷第 1 期,第 306-335 页,2024 年 3 月。 摘要本文研究了求解一类非凸两阶段随机程序的分解方法,其中第二阶段问题的目标和约束均由第一阶段变量非线性参数化。由于所得到的非凸求助函数的克拉克正则性失效,经典的分解方法,如本德斯分解和基于(增强)拉格朗日的算法,不能直接用于解决此类模型。通过探索求助函数的隐含凸凹结构,我们引入了一种基于所谓部分莫罗包络的新型分解框架。该算法根据第二阶段凸子问题的解,连续生成追索函数的强凸二次近似值,并将其添加到第一阶段主问题中。对于固定数量的方案和顺序内部采样策略,均已确定收敛性。通过数值实验证明了所提算法的有效性。
{"title":"A Decomposition Algorithm for Two-Stage Stochastic Programs with Nonconvex Recourse Functions","authors":"Hanyang Li, Ying Cui","doi":"10.1137/22m1488533","DOIUrl":"https://doi.org/10.1137/22m1488533","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 306-335, March 2024. <br/> Abstract. In this paper, we have studied a decomposition method for solving a class of nonconvex two-stage stochastic programs, where both the objective and constraints of the second-stage problem are nonlinearly parameterized by the first-stage variables. Due to the failure of the Clarke regularity of the resulting nonconvex recourse function, classical decomposition approaches such as Benders decomposition and (augmented) Lagrangian-based algorithms cannot be directly generalized to solve such models. By exploring an implicitly convex-concave structure of the recourse function, we introduce a novel decomposition framework based on the so-called partial Moreau envelope. The algorithm successively generates strongly convex quadratic approximations of the recourse function based on the solutions of the second-stage convex subproblems and adds them to the first-stage master problem. Convergence has been established for both a fixed number of scenarios and a sequential internal sampling strategy. Numerical experiments are conducted to demonstrate the effectiveness of the proposed algorithm.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"1 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139495101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sharper Bounds for Proximal Gradient Algorithms with Errors 有误差的近端梯度算法的更清晰边界
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-01-19 DOI: 10.1137/22m1480161
Anis Hamadouche, Yun Wu, Andrew M. Wallace, João F. C. Mota
SIAM Journal on Optimization, Volume 34, Issue 1, Page 278-305, March 2024.
Abstract. We analyze the convergence of the proximal gradient algorithm for convex composite problems in the presence of gradient and proximal computational inaccuracies. We generalize the deterministic analysis to the quasi-Fejér case and quantify the uncertainty incurred from approximate computing and early termination errors. We propose new probabilistic tighter bounds that we use to verify a simulated Model Predictive Control (MPC) with sparse controls problem solved with early termination, reduced precision, and proximal errors. We also show how the probabilistic bounds are more suitable than the deterministic ones for algorithm verification and more accurate for application performance guarantees. Under mild statistical assumptions, we also prove that some cumulative error terms follow a martingale property. And conforming to observations, e.g., in [M. Schmidt, N. L. Roux, and F. R. Bach, Convergence rates of inexact proximal-gradient methods for convex optimization, in Advances in Neural Information Processing Systems, 2011, pp. 1458–1466], we also show how the acceleration of the algorithm amplifies the gradient and proximal computational errors.
SIAM 优化期刊》,第 34 卷,第 1 期,第 278-305 页,2024 年 3 月。 摘要我们分析了在梯度和近端计算不准确的情况下凸复合问题的近端梯度算法的收敛性。我们将确定性分析推广到准 Fejér 情况,并量化了近似计算和提前终止误差带来的不确定性。我们提出了新的概率紧缩边界,并用它来验证一个带有稀疏控制的模拟模型预测控制 (MPC) 问题,该问题在解决时出现了提前终止、精度降低和近似误差。我们还展示了概率边界如何比确定边界更适合算法验证,以及如何更准确地保证应用性能。在温和的统计假设条件下,我们还证明了某些累积误差项遵循马氏特性。这也符合 [M. Schmidt, N. L. Rouge] 等人的观察。Schmidt, N. L. Roux, and F. R. Bach, Convergence rates of inexact proximal-gradient methods for convex optimization, in Advances in Neural Information Processing Systems, 2011, pp.
{"title":"Sharper Bounds for Proximal Gradient Algorithms with Errors","authors":"Anis Hamadouche, Yun Wu, Andrew M. Wallace, João F. C. Mota","doi":"10.1137/22m1480161","DOIUrl":"https://doi.org/10.1137/22m1480161","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 1, Page 278-305, March 2024. <br/> Abstract. We analyze the convergence of the proximal gradient algorithm for convex composite problems in the presence of gradient and proximal computational inaccuracies. We generalize the deterministic analysis to the quasi-Fejér case and quantify the uncertainty incurred from approximate computing and early termination errors. We propose new probabilistic tighter bounds that we use to verify a simulated Model Predictive Control (MPC) with sparse controls problem solved with early termination, reduced precision, and proximal errors. We also show how the probabilistic bounds are more suitable than the deterministic ones for algorithm verification and more accurate for application performance guarantees. Under mild statistical assumptions, we also prove that some cumulative error terms follow a martingale property. And conforming to observations, e.g., in [M. Schmidt, N. L. Roux, and F. R. Bach, Convergence rates of inexact proximal-gradient methods for convex optimization, in Advances in Neural Information Processing Systems, 2011, pp. 1458–1466], we also show how the acceleration of the algorithm amplifies the gradient and proximal computational errors.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":"27 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139495041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
SIAM Journal on Optimization
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1